var/home/core/zuul-output/0000755000175000017500000000000015070113136014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070124713015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004702606615070124705017711 0ustar rootrootOct 04 03:40:24 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 03:40:24 crc restorecon[4671]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:24 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 03:40:25 crc restorecon[4671]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 03:40:26 crc kubenswrapper[4726]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.239901 4726 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246180 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246218 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246227 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246236 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246243 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246251 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246260 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246267 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246273 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246279 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246285 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246293 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246299 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246306 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246312 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246318 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246324 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246330 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246335 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246341 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246347 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246353 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246358 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246364 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246369 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246374 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246381 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246387 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246392 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246404 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246409 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246414 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246419 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246424 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246429 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246434 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246440 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246447 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246453 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246460 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246465 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246470 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246475 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246479 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246485 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246490 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246496 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246501 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246506 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246511 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246516 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246522 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246527 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246533 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246538 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246547 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246553 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246559 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246565 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246570 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246575 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246580 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246585 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246590 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246594 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246601 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246607 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246612 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246618 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246623 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.246628 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247812 4726 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247828 4726 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247838 4726 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247845 4726 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247853 4726 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247860 4726 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247872 4726 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247879 4726 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247885 4726 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247890 4726 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247899 4726 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247905 4726 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247911 4726 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247917 4726 flags.go:64] FLAG: --cgroup-root="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247923 4726 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247928 4726 flags.go:64] FLAG: --client-ca-file="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247936 4726 flags.go:64] FLAG: --cloud-config="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247942 4726 flags.go:64] FLAG: --cloud-provider="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247948 4726 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247958 4726 flags.go:64] FLAG: --cluster-domain="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247964 4726 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247970 4726 flags.go:64] FLAG: --config-dir="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247976 4726 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247982 4726 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247989 4726 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.247995 4726 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248002 4726 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248008 4726 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248014 4726 flags.go:64] FLAG: --contention-profiling="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248020 4726 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248026 4726 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248032 4726 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248037 4726 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248045 4726 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248051 4726 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248056 4726 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248062 4726 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248068 4726 flags.go:64] FLAG: --enable-server="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248074 4726 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248081 4726 flags.go:64] FLAG: --event-burst="100" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248087 4726 flags.go:64] FLAG: --event-qps="50" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248093 4726 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248099 4726 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248123 4726 flags.go:64] FLAG: --eviction-hard="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248131 4726 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248138 4726 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248144 4726 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248150 4726 flags.go:64] FLAG: --eviction-soft="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248158 4726 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248163 4726 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248169 4726 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248175 4726 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248181 4726 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248186 4726 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248192 4726 flags.go:64] FLAG: --feature-gates="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248199 4726 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248205 4726 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248211 4726 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248217 4726 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248223 4726 flags.go:64] FLAG: --healthz-port="10248" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248229 4726 flags.go:64] FLAG: --help="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248235 4726 flags.go:64] FLAG: --hostname-override="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248241 4726 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248249 4726 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248255 4726 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248260 4726 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248266 4726 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248272 4726 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248278 4726 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248284 4726 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248290 4726 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248295 4726 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248302 4726 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248308 4726 flags.go:64] FLAG: --kube-reserved="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248313 4726 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248319 4726 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248325 4726 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248331 4726 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248338 4726 flags.go:64] FLAG: --lock-file="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248343 4726 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248351 4726 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248357 4726 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248372 4726 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248379 4726 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248385 4726 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248391 4726 flags.go:64] FLAG: --logging-format="text" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248397 4726 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248404 4726 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248410 4726 flags.go:64] FLAG: --manifest-url="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248416 4726 flags.go:64] FLAG: --manifest-url-header="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248423 4726 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248429 4726 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248437 4726 flags.go:64] FLAG: --max-pods="110" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248445 4726 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248452 4726 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248458 4726 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248464 4726 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248470 4726 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248476 4726 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248483 4726 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248498 4726 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248504 4726 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248510 4726 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248516 4726 flags.go:64] FLAG: --pod-cidr="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248522 4726 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248531 4726 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248537 4726 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248543 4726 flags.go:64] FLAG: --pods-per-core="0" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248549 4726 flags.go:64] FLAG: --port="10250" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248555 4726 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248560 4726 flags.go:64] FLAG: --provider-id="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248567 4726 flags.go:64] FLAG: --qos-reserved="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248572 4726 flags.go:64] FLAG: --read-only-port="10255" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248581 4726 flags.go:64] FLAG: --register-node="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248588 4726 flags.go:64] FLAG: --register-schedulable="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248594 4726 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248608 4726 flags.go:64] FLAG: --registry-burst="10" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248614 4726 flags.go:64] FLAG: --registry-qps="5" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248623 4726 flags.go:64] FLAG: --reserved-cpus="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248630 4726 flags.go:64] FLAG: --reserved-memory="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248640 4726 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248646 4726 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248654 4726 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248661 4726 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248668 4726 flags.go:64] FLAG: --runonce="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248676 4726 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248685 4726 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248693 4726 flags.go:64] FLAG: --seccomp-default="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248700 4726 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248707 4726 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248715 4726 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248721 4726 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248729 4726 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248736 4726 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248741 4726 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248747 4726 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248753 4726 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248759 4726 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248765 4726 flags.go:64] FLAG: --system-cgroups="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248771 4726 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248780 4726 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248785 4726 flags.go:64] FLAG: --tls-cert-file="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248791 4726 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248800 4726 flags.go:64] FLAG: --tls-min-version="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248806 4726 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248816 4726 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248822 4726 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248828 4726 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248834 4726 flags.go:64] FLAG: --v="2" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248842 4726 flags.go:64] FLAG: --version="false" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248850 4726 flags.go:64] FLAG: --vmodule="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248857 4726 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.248863 4726 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249002 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249010 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249016 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249021 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249027 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249033 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249038 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249044 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249049 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249054 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249059 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249064 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249069 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249074 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249079 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249083 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249088 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249093 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249098 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249126 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249132 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249137 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249142 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249147 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249155 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249160 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249165 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249169 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249174 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249179 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249184 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249189 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249194 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249199 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249204 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249209 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249214 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249219 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249248 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249255 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249262 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249268 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249274 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249280 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249286 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249292 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249297 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249302 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249307 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249311 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249316 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249323 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249329 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249335 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249340 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249345 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249352 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249357 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249364 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249370 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249375 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249380 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249386 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249391 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249396 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249401 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249406 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249411 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249416 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249421 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.249426 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.249442 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.258621 4726 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.258653 4726 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258793 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258805 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258815 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258823 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258833 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258841 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258849 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258856 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258864 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258872 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258880 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258887 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258895 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258903 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258911 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258922 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258935 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258944 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258952 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258961 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258968 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258976 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258984 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.258991 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259000 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259008 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259016 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259024 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259031 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259040 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259048 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259056 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259064 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259071 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259081 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259089 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259126 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259137 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259145 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259154 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259162 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259170 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259178 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259186 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259194 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259202 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259209 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259219 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259228 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259237 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259245 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259254 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259262 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259269 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259277 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259285 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259293 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259300 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259308 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259316 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259323 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259332 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259339 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259349 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259357 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259368 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259378 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259386 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259394 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259403 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259412 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.259426 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259682 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259696 4726 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259705 4726 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259713 4726 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259722 4726 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259732 4726 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259743 4726 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259751 4726 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259761 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259771 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259779 4726 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259786 4726 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259794 4726 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259802 4726 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259810 4726 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259817 4726 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259825 4726 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259832 4726 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259840 4726 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259848 4726 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259856 4726 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259863 4726 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259871 4726 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259879 4726 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259887 4726 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259896 4726 feature_gate.go:330] unrecognized feature gate: Example Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259904 4726 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259912 4726 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259919 4726 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259928 4726 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259936 4726 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259944 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259951 4726 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259959 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259967 4726 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259976 4726 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259983 4726 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259991 4726 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.259999 4726 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260006 4726 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260014 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260022 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260030 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260037 4726 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260045 4726 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260052 4726 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260060 4726 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260068 4726 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260075 4726 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260083 4726 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260091 4726 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260121 4726 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260132 4726 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260142 4726 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260150 4726 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260159 4726 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260168 4726 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260175 4726 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260186 4726 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260195 4726 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260203 4726 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260213 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260222 4726 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260232 4726 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260242 4726 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260250 4726 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260258 4726 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260267 4726 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260276 4726 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260284 4726 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.260299 4726 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.260311 4726 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.261382 4726 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.266861 4726 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.266972 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.268905 4726 server.go:997] "Starting client certificate rotation" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.268944 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.269970 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 21:29:17.988410655 +0000 UTC Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.270090 4726 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1625h48m51.718325771s for next certificate rotation Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.297747 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.303440 4726 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.329758 4726 log.go:25] "Validated CRI v1 runtime API" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.370336 4726 log.go:25] "Validated CRI v1 image API" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.372918 4726 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.381277 4726 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-03-36-02-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.381340 4726 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.413449 4726 manager.go:217] Machine: {Timestamp:2025-10-04 03:40:26.409652916 +0000 UTC m=+0.584276209 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:64e1ddb6-4f16-4c65-97b7-a8481b56f368 BootID:10cbd887-f465-43af-a376-e5f244cebbdc Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ec:73:d2 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ec:73:d2 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:3a:bf:6a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:dc:c9:ee Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:99:db:b3 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b9:d3:65 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:f6:a8:bb:8f:b6 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:26:8f:19:30:0b:dd Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.413910 4726 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.414102 4726 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.416946 4726 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.417377 4726 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.417441 4726 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.417846 4726 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.417867 4726 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.418592 4726 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.418654 4726 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.418917 4726 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.419053 4726 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.422455 4726 kubelet.go:418] "Attempting to sync node with API server" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.422492 4726 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.422535 4726 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.422556 4726 kubelet.go:324] "Adding apiserver pod source" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.422574 4726 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.427374 4726 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.428616 4726 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.430565 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.430667 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.430734 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.430776 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.431251 4726 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435281 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435350 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435365 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435379 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435400 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435447 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435462 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435484 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435500 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435515 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435574 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.435588 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.437271 4726 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.438280 4726 server.go:1280] "Started kubelet" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.438466 4726 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.438669 4726 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.439295 4726 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.440154 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:26 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.441635 4726 server.go:460] "Adding debug handlers to kubelet server" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.442684 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.442722 4726 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.442798 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 05:14:23.819143289 +0000 UTC Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.442839 4726 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2041h33m57.376309846s for next certificate rotation Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.443316 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.443331 4726 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.443650 4726 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.443677 4726 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.444399 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.444431 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.176:6443: connect: connection refused" interval="200ms" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.444487 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.445662 4726 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.445706 4726 factory.go:55] Registering systemd factory Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.445726 4726 factory.go:221] Registration of the systemd container factory successfully Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.446292 4726 factory.go:153] Registering CRI-O factory Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.446332 4726 factory.go:221] Registration of the crio container factory successfully Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.446370 4726 factory.go:103] Registering Raw factory Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.446394 4726 manager.go:1196] Started watching for new ooms in manager Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.447590 4726 manager.go:319] Starting recovery of all containers Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.446240 4726 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.176:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2cb2eeaa1256 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 03:40:26.43823471 +0000 UTC m=+0.612857963,LastTimestamp:2025-10-04 03:40:26.43823471 +0000 UTC m=+0.612857963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455250 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455332 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455356 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455381 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455401 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455418 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455441 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455460 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455492 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455510 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455535 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455565 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455582 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455613 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455638 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455657 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455674 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455697 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455715 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.455736 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457777 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457845 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457863 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457885 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457895 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457913 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457929 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457948 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.457969 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458013 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458026 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458042 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458054 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458071 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458083 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458098 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458197 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458212 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458224 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458239 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458256 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458273 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458284 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458298 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458313 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458326 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458343 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458357 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458375 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458393 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458408 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458433 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458453 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458477 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458499 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458520 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458542 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458559 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458581 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458596 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458610 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458626 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458636 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458919 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458944 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458961 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.458987 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.459001 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.459019 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.459035 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.463970 4726 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464256 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464298 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464328 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464349 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464370 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464393 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464425 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464455 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464481 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464507 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464533 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464555 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464595 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464622 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464645 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464670 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464693 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464717 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464738 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464756 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464789 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464809 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464837 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464861 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464878 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464902 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464931 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464950 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464976 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.464995 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.465023 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.465043 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.465922 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.465957 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466009 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466035 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466058 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466081 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466103 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466147 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466169 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466189 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466209 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466227 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466282 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466300 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466317 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466333 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466352 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466369 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466385 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466404 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466423 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466444 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466460 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466479 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466495 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466513 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466529 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466545 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466562 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466581 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466599 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466615 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466632 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466651 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466669 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466686 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466702 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466718 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466737 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466753 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466769 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466788 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466803 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466818 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466833 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466851 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466866 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466883 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466899 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466915 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466933 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466951 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466970 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.466987 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467003 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467021 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467040 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467059 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467074 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467091 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467157 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467176 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467196 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467214 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467231 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467300 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467320 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467337 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467356 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467372 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467390 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467407 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467433 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467452 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467471 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467489 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467505 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467523 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467541 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467557 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467573 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467589 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467603 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467620 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467635 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467654 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467669 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467688 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467704 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467720 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467752 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467769 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467784 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467798 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467812 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467827 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467843 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467859 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467875 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467892 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467910 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467927 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467943 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467960 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467977 4726 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.467992 4726 reconstruct.go:97] "Volume reconstruction finished" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.468006 4726 reconciler.go:26] "Reconciler: start to sync state" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.483954 4726 manager.go:324] Recovery completed Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.498256 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.501193 4726 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.501237 4726 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.501272 4726 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.501411 4726 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 03:40:26 crc kubenswrapper[4726]: W1004 03:40:26.502611 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.502680 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.507058 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.508919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.509051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.509166 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.509995 4726 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.510014 4726 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.510035 4726 state_mem.go:36] "Initialized new in-memory state store" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.539478 4726 policy_none.go:49] "None policy: Start" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.540583 4726 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.540617 4726 state_mem.go:35] "Initializing new in-memory state store" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.543992 4726 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.601988 4726 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.613852 4726 manager.go:334] "Starting Device Plugin manager" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.613924 4726 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.613946 4726 server.go:79] "Starting device plugin registration server" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.614651 4726 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.614680 4726 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.615705 4726 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.615868 4726 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.615891 4726 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.629557 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.645421 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.176:6443: connect: connection refused" interval="400ms" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.715524 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.717243 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.717323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.717349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.717401 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.718081 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.176:6443: connect: connection refused" node="crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.802745 4726 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.802932 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.806290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.806346 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.806364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.806650 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.806964 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.807050 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808877 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.808835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.809131 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.809223 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.810266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.810340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.810355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.810586 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.810963 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.811280 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.812306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.812344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.812360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.812491 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.813323 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.813368 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814291 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814322 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.814956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.815467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.815499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.815512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.815977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.816036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.816054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.825617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.825653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.825667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.872906 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.872950 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.872976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.873004 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.873024 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.873049 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.873337 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.879218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.879333 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.879424 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.879481 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.879592 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.880268 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.880349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.880395 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.918516 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.919886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.919925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.919937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.919963 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:26 crc kubenswrapper[4726]: E1004 03:40:26.920503 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.176:6443: connect: connection refused" node="crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981304 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981403 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981430 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981449 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981521 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981543 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981585 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981511 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981644 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981703 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981608 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981744 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981789 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981825 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981836 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981664 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981870 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981865 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981938 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.981962 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:26 crc kubenswrapper[4726]: I1004 03:40:26.982060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.046928 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.176:6443: connect: connection refused" interval="800ms" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.150835 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.155576 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.170085 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.196435 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.204799 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-213d8e37f20bed1b39cfb31abe722612be013f3397b6c7b186b52a5fa33cb1a4 WatchSource:0}: Error finding container 213d8e37f20bed1b39cfb31abe722612be013f3397b6c7b186b52a5fa33cb1a4: Status 404 returned error can't find the container with id 213d8e37f20bed1b39cfb31abe722612be013f3397b6c7b186b52a5fa33cb1a4 Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.206151 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.206834 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d8e871c0dda60fdee6192270f15c073ba84959f68e5ef644918d54b233b91274 WatchSource:0}: Error finding container d8e871c0dda60fdee6192270f15c073ba84959f68e5ef644918d54b233b91274: Status 404 returned error can't find the container with id d8e871c0dda60fdee6192270f15c073ba84959f68e5ef644918d54b233b91274 Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.213481 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6e4bebb8f6409fe1ab6a22f53617108f9b4ea9fd35d4793f01a343345daf1912 WatchSource:0}: Error finding container 6e4bebb8f6409fe1ab6a22f53617108f9b4ea9fd35d4793f01a343345daf1912: Status 404 returned error can't find the container with id 6e4bebb8f6409fe1ab6a22f53617108f9b4ea9fd35d4793f01a343345daf1912 Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.220664 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d140eb047433082a8d825281925ba1486066b27e9b23f10eef5001b404ad5d4d WatchSource:0}: Error finding container d140eb047433082a8d825281925ba1486066b27e9b23f10eef5001b404ad5d4d: Status 404 returned error can't find the container with id d140eb047433082a8d825281925ba1486066b27e9b23f10eef5001b404ad5d4d Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.227069 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b470901ff5dbc8e5d473299d97e27cd9442968017b1a29327769cc953cf5003d WatchSource:0}: Error finding container b470901ff5dbc8e5d473299d97e27cd9442968017b1a29327769cc953cf5003d: Status 404 returned error can't find the container with id b470901ff5dbc8e5d473299d97e27cd9442968017b1a29327769cc953cf5003d Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.321088 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.322505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.322653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.322685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.322784 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.323704 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.176:6443: connect: connection refused" node="crc" Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.335384 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.335450 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.441773 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.506179 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"213d8e37f20bed1b39cfb31abe722612be013f3397b6c7b186b52a5fa33cb1a4"} Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.507482 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b470901ff5dbc8e5d473299d97e27cd9442968017b1a29327769cc953cf5003d"} Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.509569 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d140eb047433082a8d825281925ba1486066b27e9b23f10eef5001b404ad5d4d"} Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.510474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6e4bebb8f6409fe1ab6a22f53617108f9b4ea9fd35d4793f01a343345daf1912"} Oct 04 03:40:27 crc kubenswrapper[4726]: I1004 03:40:27.511469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d8e871c0dda60fdee6192270f15c073ba84959f68e5ef644918d54b233b91274"} Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.720682 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.720825 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.762065 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.762198 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.847672 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.176:6443: connect: connection refused" interval="1.6s" Oct 04 03:40:27 crc kubenswrapper[4726]: W1004 03:40:27.923754 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:27 crc kubenswrapper[4726]: E1004 03:40:27.924171 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.124638 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.126629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.126693 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.126713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.126752 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:28 crc kubenswrapper[4726]: E1004 03:40:28.127460 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.176:6443: connect: connection refused" node="crc" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.441518 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.517917 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.517994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.520554 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e9951597ad325756b10cd27777bfe38a09befc91ac692074d9a53c1c0709b9a9" exitCode=0 Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.520641 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e9951597ad325756b10cd27777bfe38a09befc91ac692074d9a53c1c0709b9a9"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.520781 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.522380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.522411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.522423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.522835 4726 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998" exitCode=0 Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.522979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.523073 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.524258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.524299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.524317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.525512 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c" exitCode=0 Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.525602 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.525622 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.527465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.527523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.527549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.528456 4726 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162" exitCode=0 Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.528530 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.528508 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162"} Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.529483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.529544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.529562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.529578 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.530586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.530618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:28 crc kubenswrapper[4726]: I1004 03:40:28.530634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: W1004 03:40:29.050612 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:29 crc kubenswrapper[4726]: E1004 03:40:29.050704 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.441009 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:29 crc kubenswrapper[4726]: E1004 03:40:29.449179 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.176:6443: connect: connection refused" interval="3.2s" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.535343 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.535386 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.535395 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.535404 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.537760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.537821 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.537837 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.537865 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.538971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.539005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.539015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.543664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.543726 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.543681 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548652 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="27fe1952f17df5b8c20aaf2e7d7e7218e7191f884d60540c33e31c4b2e4a3667" exitCode=0 Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548718 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"27fe1952f17df5b8c20aaf2e7d7e7218e7191f884d60540c33e31c4b2e4a3667"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.548733 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.549368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.549390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.549400 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.551017 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b79980747936da340dcc0b0fecc9d02f5802535b81766bcdb254ecbfb44fc835"} Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.551099 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.552490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.552516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.552529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: W1004 03:40:29.574756 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:29 crc kubenswrapper[4726]: E1004 03:40:29.574827 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:29 crc kubenswrapper[4726]: W1004 03:40:29.587759 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.176:6443: connect: connection refused Oct 04 03:40:29 crc kubenswrapper[4726]: E1004 03:40:29.587813 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.176:6443: connect: connection refused" logger="UnhandledError" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.649376 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.728589 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.729860 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.729898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.729908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:29 crc kubenswrapper[4726]: I1004 03:40:29.729934 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:29 crc kubenswrapper[4726]: E1004 03:40:29.730341 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.176:6443: connect: connection refused" node="crc" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.557613 4726 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c89499a5e82dcd2a9a0e12a814f299455f687b48d16bf70dabe7ac55647cc588" exitCode=0 Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.557748 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c89499a5e82dcd2a9a0e12a814f299455f687b48d16bf70dabe7ac55647cc588"} Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.557811 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.559155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.559214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.559235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.563435 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310"} Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.563537 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.563617 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.563832 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.564326 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.564516 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.564983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.565031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.565048 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.565616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.565652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.565670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.580871 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.580960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.580984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.581414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.581597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:30 crc kubenswrapper[4726]: I1004 03:40:30.581768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586395 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586407 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"db14885bfc80d4bb16a6b9543f86db53a4705be1be2342a366b97e2bf95556bb"} Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586450 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586465 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"26e575b3f89314e4752c49f999b56424adda9601fbc6f48275011c071ad2929e"} Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586517 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.586543 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587934 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:31 crc kubenswrapper[4726]: I1004 03:40:31.587945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.379734 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.597710 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8716c5f2a510792dc941290d2314aa9ece036b5cd4a9f4c54b968895ba62f630"} Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.597784 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32f51ef5cfad2fadb6af4accbfe5833046212b04bb4289e30395f271d352fd42"} Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.597809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9ca16cc409413075d7708b56da5d7e3df2c4fbda8413687ae62c77b47674a578"} Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.597823 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.597726 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.598048 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.599784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.930922 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.932632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.932692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.932713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:32 crc kubenswrapper[4726]: I1004 03:40:32.932750 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.462193 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.462429 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.464158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.464194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.464208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.579250 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.600946 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.602340 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.602391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:33 crc kubenswrapper[4726]: I1004 03:40:33.602410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.603761 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.606368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.606435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.606454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.861248 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.861670 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.863476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.863575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:34 crc kubenswrapper[4726]: I1004 03:40:34.863599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.039231 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.518756 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.607018 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.607018 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.608987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.808629 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.808964 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.811298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.811399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.811418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:35 crc kubenswrapper[4726]: I1004 03:40:35.815354 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.462770 4726 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.462880 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.610188 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.613159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.613218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:36 crc kubenswrapper[4726]: I1004 03:40:36.613235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:36 crc kubenswrapper[4726]: E1004 03:40:36.630511 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.559897 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.612543 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.613900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.613957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.613975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:37 crc kubenswrapper[4726]: I1004 03:40:37.618586 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:38 crc kubenswrapper[4726]: I1004 03:40:38.615646 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:38 crc kubenswrapper[4726]: I1004 03:40:38.616803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:38 crc kubenswrapper[4726]: I1004 03:40:38.616850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:38 crc kubenswrapper[4726]: I1004 03:40:38.616868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:40 crc kubenswrapper[4726]: W1004 03:40:40.328529 4726 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.328671 4726 trace.go:236] Trace[822152713]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:40:30.326) (total time: 10002ms): Oct 04 03:40:40 crc kubenswrapper[4726]: Trace[822152713]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:40:40.328) Oct 04 03:40:40 crc kubenswrapper[4726]: Trace[822152713]: [10.002005422s] [10.002005422s] END Oct 04 03:40:40 crc kubenswrapper[4726]: E1004 03:40:40.328715 4726 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.442011 4726 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.622812 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.624964 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310" exitCode=255 Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.625009 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310"} Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.625181 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.626309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.626390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.626413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:40 crc kubenswrapper[4726]: I1004 03:40:40.627621 4726 scope.go:117] "RemoveContainer" containerID="e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.628763 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.632487 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa"} Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.632752 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.634011 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.634055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.634069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.677662 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.677763 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.693342 4726 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 04 03:40:41 crc kubenswrapper[4726]: I1004 03:40:41.693415 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.643011 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.643312 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.644722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.644767 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.644784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:43 crc kubenswrapper[4726]: I1004 03:40:43.665355 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.641471 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.643004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.643062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.643081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.862031 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.862358 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.864194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.864272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:44 crc kubenswrapper[4726]: I1004 03:40:44.864298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.048961 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.644470 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.645787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.645838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.645850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:45 crc kubenswrapper[4726]: I1004 03:40:45.651836 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.462949 4726 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.463083 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:40:46 crc kubenswrapper[4726]: E1004 03:40:46.631185 4726 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.646816 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.648190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.648253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.648275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:46 crc kubenswrapper[4726]: E1004 03:40:46.678101 4726 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.683386 4726 trace.go:236] Trace[1878038035]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:40:33.238) (total time: 13444ms): Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[1878038035]: ---"Objects listed" error: 13444ms (03:40:46.683) Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[1878038035]: [13.44495969s] [13.44495969s] END Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.683425 4726 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.684400 4726 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.684572 4726 trace.go:236] Trace[584486836]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:40:34.654) (total time: 12029ms): Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[584486836]: ---"Objects listed" error: 12029ms (03:40:46.684) Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[584486836]: [12.029575125s] [12.029575125s] END Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.684605 4726 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 03:40:46 crc kubenswrapper[4726]: E1004 03:40:46.688696 4726 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.689811 4726 trace.go:236] Trace[1337064026]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 03:40:34.158) (total time: 12530ms): Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[1337064026]: ---"Objects listed" error: 12530ms (03:40:46.689) Oct 04 03:40:46 crc kubenswrapper[4726]: Trace[1337064026]: [12.530890896s] [12.530890896s] END Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.690098 4726 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 03:40:46 crc kubenswrapper[4726]: I1004 03:40:46.726048 4726 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.435218 4726 apiserver.go:52] "Watching apiserver" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.440420 4726 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.440705 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441220 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441261 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.441432 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441461 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441509 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.441668 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.441771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.441825 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.445041 4726 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.446952 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.446976 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.446982 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.447422 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.447621 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.449582 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.449955 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.449962 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.452355 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488319 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488363 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488382 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488403 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488422 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488446 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488463 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488479 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488495 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488511 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488528 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488544 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488559 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488573 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488590 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488605 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488618 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488649 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488684 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488717 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488754 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488774 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488793 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488810 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488832 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488857 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488881 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488928 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488953 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488972 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.488992 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489016 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489037 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489060 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489086 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489125 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489128 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489149 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489169 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489243 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489329 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489502 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489696 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489820 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489956 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.489991 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490166 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490198 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490242 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490263 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490305 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492968 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493018 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493079 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493156 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493186 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493210 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493232 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493257 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493300 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493321 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493343 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493386 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495007 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495065 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495138 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495172 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495223 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495255 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495285 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495347 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495383 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495415 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495449 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495523 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495557 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495589 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495620 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495668 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495755 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495788 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495833 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495868 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495902 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495952 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495999 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496031 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496087 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490331 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490436 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490592 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490636 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490746 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490851 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490949 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490954 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490969 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.490994 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491153 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491669 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491741 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491878 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.491973 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492012 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492080 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492227 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492434 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.492909 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493569 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493707 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.493939 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494163 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494298 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494311 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494677 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.494685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495158 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495372 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495398 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495405 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495404 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495419 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495737 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495838 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495863 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.495870 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496015 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496252 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496260 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496311 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497098 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497402 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.496722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497591 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497595 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497617 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497644 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497665 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497729 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497758 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497785 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497797 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497814 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497824 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497842 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497868 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497895 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497919 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497948 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497978 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.497978 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498002 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498252 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498285 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498453 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498493 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498531 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498581 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498591 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498616 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498649 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498687 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498768 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498816 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498849 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498901 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498939 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498976 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499034 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499071 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499237 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499273 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499393 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499459 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499493 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499554 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499590 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499740 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499808 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499840 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499882 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500791 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500826 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500857 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500888 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500920 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500954 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500988 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501162 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501200 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501270 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501304 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501337 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501369 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501401 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501435 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501502 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501533 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501566 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501601 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501670 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501702 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501735 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501767 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501834 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501866 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501898 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501930 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501988 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502024 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502055 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502182 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502288 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502327 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502360 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502398 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502437 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502477 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502514 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502551 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502583 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502617 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502651 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502686 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502720 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502809 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502897 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503363 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503420 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503465 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503543 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503713 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503824 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504023 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504240 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504334 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504358 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504380 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504400 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504421 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504447 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504476 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504504 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504536 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504693 4726 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504805 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504839 4726 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504867 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504894 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504922 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504950 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504996 4726 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505029 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505057 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505100 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505159 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505189 4726 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505219 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505247 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505274 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505301 4726 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505328 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505356 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505384 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505412 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505440 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505468 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505498 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505525 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505553 4726 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505581 4726 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507574 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507627 4726 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507773 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507805 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507832 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507857 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507884 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507913 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507944 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507972 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508002 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508031 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508058 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508086 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508147 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508178 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508208 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508235 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508262 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508289 4726 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508316 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508343 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508369 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508396 4726 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508425 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508454 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508481 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508510 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508538 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508566 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508598 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508624 4726 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508651 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508679 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508706 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508733 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508760 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.498895 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499129 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499295 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499398 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499677 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499744 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.510059 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.499821 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500145 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500277 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500389 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500691 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.500723 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.501558 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502125 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502798 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503074 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503147 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503212 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503308 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503474 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503667 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503798 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504261 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.503623 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504651 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504787 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504852 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.504964 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505026 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505056 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505256 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.502759 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505421 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505462 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505455 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505506 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505571 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505825 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505843 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.505986 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506041 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506234 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506352 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506476 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506620 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.506893 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507675 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.507775 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.508840 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.509045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.509736 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.509781 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.511022 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.511673 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.512015 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.512021 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.512541 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.512952 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.513069 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.513369 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.513515 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.513594 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.513994 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514401 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514557 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514633 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514918 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.514955 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.515181 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.515326 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.515511 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.516165 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.516350 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.516540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.516808 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517141 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517258 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.517667 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:40:48.017638464 +0000 UTC m=+22.192261787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517722 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517814 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517934 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517982 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.518195 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.518365 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:48.018351353 +0000 UTC m=+22.192974666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.518784 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.518799 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.519522 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.519802 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.520051 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.520236 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.520656 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.520817 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.521071 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.521417 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.521685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.522794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.523488 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.524197 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.525091 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.529745 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.529814 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.530056 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.530119 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.517813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.518286 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.530407 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.530703 4726 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.531035 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.532814 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.532922 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:48.032899066 +0000 UTC m=+22.207522279 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.533310 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.533552 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.533776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.534427 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.535685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.537261 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.537765 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.537799 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.537810 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.537925 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:48.037851626 +0000 UTC m=+22.212474839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.541886 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.542844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.542863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.542935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.542973 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.543001 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.543015 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.543345 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:48.043079873 +0000 UTC m=+22.217703086 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.543443 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.548482 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.549498 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.562616 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.565647 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.568299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.575307 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.578805 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.586782 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.594875 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.602451 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609291 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609348 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609361 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609370 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609378 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609388 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609397 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609405 4726 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609413 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609421 4726 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609429 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609438 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609447 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609456 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609464 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609473 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609481 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609490 4726 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609499 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609508 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609517 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609525 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609534 4726 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609543 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609552 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609560 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609569 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609578 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609585 4726 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609595 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609603 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609610 4726 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609618 4726 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609626 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609634 4726 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609642 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609650 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609658 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609666 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609686 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609705 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609715 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609726 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609736 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609745 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609755 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609765 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609774 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609784 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609793 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609800 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609809 4726 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609818 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609826 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609835 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609827 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609843 4726 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609899 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609912 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609923 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609933 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609943 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609954 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609963 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609973 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609983 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.609992 4726 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610014 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610029 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610040 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610050 4726 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610058 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610019 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610129 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610141 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610150 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610158 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610167 4726 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610174 4726 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610182 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610192 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610200 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610209 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610217 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610224 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610232 4726 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610240 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610256 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610264 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610272 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610280 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610364 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610377 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610388 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610398 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610410 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610419 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610427 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610437 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610445 4726 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610453 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610460 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610469 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610493 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610501 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610512 4726 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610520 4726 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610527 4726 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610537 4726 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610546 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610554 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610564 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610572 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610580 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610588 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610596 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610605 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610612 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610621 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610629 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610637 4726 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610645 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.610653 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.650642 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.650968 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.652466 4726 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa" exitCode=255 Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.652498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa"} Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.652539 4726 scope.go:117] "RemoveContainer" containerID="e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.666033 4726 scope.go:117] "RemoveContainer" containerID="7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa" Oct 04 03:40:47 crc kubenswrapper[4726]: E1004 03:40:47.666245 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.667514 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.667750 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.676493 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.685862 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.697274 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.707704 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.719139 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.758479 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.776498 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 03:40:47 crc kubenswrapper[4726]: I1004 03:40:47.781834 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 03:40:47 crc kubenswrapper[4726]: W1004 03:40:47.784686 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-68e02cbca47c54e5fc4faab0c4c6f21517f0f5b6354b826e44ce1f79e3efbc1f WatchSource:0}: Error finding container 68e02cbca47c54e5fc4faab0c4c6f21517f0f5b6354b826e44ce1f79e3efbc1f: Status 404 returned error can't find the container with id 68e02cbca47c54e5fc4faab0c4c6f21517f0f5b6354b826e44ce1f79e3efbc1f Oct 04 03:40:47 crc kubenswrapper[4726]: W1004 03:40:47.793195 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-d20d0588c9822d6172019c8e9c5aa08a914862e30dbd7cf8d5048edb0418a53a WatchSource:0}: Error finding container d20d0588c9822d6172019c8e9c5aa08a914862e30dbd7cf8d5048edb0418a53a: Status 404 returned error can't find the container with id d20d0588c9822d6172019c8e9c5aa08a914862e30dbd7cf8d5048edb0418a53a Oct 04 03:40:47 crc kubenswrapper[4726]: W1004 03:40:47.803613 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-20d5aff6a33fd6b42b3f49b139edbff1fd364195f02a73db9ca07d2b7ccf8666 WatchSource:0}: Error finding container 20d5aff6a33fd6b42b3f49b139edbff1fd364195f02a73db9ca07d2b7ccf8666: Status 404 returned error can't find the container with id 20d5aff6a33fd6b42b3f49b139edbff1fd364195f02a73db9ca07d2b7ccf8666 Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.114456 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.114512 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.114532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.114554 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114603 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:40:49.114579303 +0000 UTC m=+23.289202516 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.114662 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114663 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114668 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114732 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114738 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114810 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114725 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:49.114719097 +0000 UTC m=+23.289342310 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114856 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:49.11483695 +0000 UTC m=+23.289460223 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114867 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:49.11486251 +0000 UTC m=+23.289485713 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114746 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114887 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114894 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.114920 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:49.114914952 +0000 UTC m=+23.289538165 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.410730 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-95wxp"] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.411143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.411685 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-snjd6"] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.412501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.413635 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.414184 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-9777l"] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.414460 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.414924 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.415585 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.415748 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.415876 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.416810 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.416870 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5c96v"] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.416977 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.417906 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-t24l4"] Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.418021 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.417194 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.420366 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.420428 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.421065 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.423226 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.423228 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.423362 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.423496 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.423703 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.425324 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.425441 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.425573 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.426711 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.426801 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.426810 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.426939 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.432259 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.445169 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.460686 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.471816 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.484237 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.494391 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:40Z\\\",\\\"message\\\":\\\"W1004 03:40:29.612852 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 03:40:29.613285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759549229 cert, and key in /tmp/serving-cert-744945107/serving-signer.crt, /tmp/serving-cert-744945107/serving-signer.key\\\\nI1004 03:40:29.965438 1 observer_polling.go:159] Starting file observer\\\\nW1004 03:40:29.972250 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 03:40:29.972510 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:29.977271 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-744945107/tls.crt::/tmp/serving-cert-744945107/tls.key\\\\\\\"\\\\nF1004 03:40:40.211393 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.504754 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.505282 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.505797 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.506977 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.507614 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.508552 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.509073 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.509758 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.510654 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.511304 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.512224 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.512721 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.513732 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.514264 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.514768 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.515665 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.516263 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.517165 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.517561 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518138 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518675 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-k8s-cni-cncf-io\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518849 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-etc-kubernetes\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-system-cni-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.518972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-os-release\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519016 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-bin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-multus\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519226 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-cnibin\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519280 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519303 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519352 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-multus-certs\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519392 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4kv8\" (UniqueName: \"kubernetes.io/projected/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-kube-api-access-f4kv8\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519438 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-binary-copy\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519498 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqlqr\" (UniqueName: \"kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519532 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cni-binary-copy\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519617 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-hostroot\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519646 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-daemon-config\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519706 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-netns\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519726 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519807 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-rootfs\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519821 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8bxc\" (UniqueName: \"kubernetes.io/projected/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-kube-api-access-s8bxc\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519837 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519866 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-system-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.519994 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-socket-dir-parent\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520027 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-proxy-tls\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520059 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-conf-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520074 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520114 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520127 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520158 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520176 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520192 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520207 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb0a03c5-46d9-4815-ab49-a93336faeff2-hosts-file\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520232 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5q2w\" (UniqueName: \"kubernetes.io/projected/bb0a03c5-46d9-4815-ab49-a93336faeff2-kube-api-access-l5q2w\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520246 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520261 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520286 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-mcd-auth-proxy-config\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520301 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfzq4\" (UniqueName: \"kubernetes.io/projected/fab2e099-b326-498e-bd92-96260472132c-kube-api-access-xfzq4\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520315 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520341 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520355 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520369 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cnibin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520382 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-os-release\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520397 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-kubelet\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520378 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520412 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.520758 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.521913 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.522446 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.524285 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.524795 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.525557 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.526855 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.527458 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.528564 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.529295 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.530262 4726 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.530385 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.532229 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.533249 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.533624 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.533770 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.535842 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.536652 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.537696 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.538509 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.539794 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.540499 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.541650 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.542369 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.543518 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.544066 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.545101 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.545702 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.546982 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.547616 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.548568 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.549130 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.550210 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.550881 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.551782 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.583985 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.592996 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.608633 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621341 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621392 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-k8s-cni-cncf-io\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621492 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621659 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-k8s-cni-cncf-io\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621878 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621941 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-etc-kubernetes\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-system-cni-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.621970 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-os-release\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622015 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-etc-kubernetes\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622045 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-system-cni-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622085 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622411 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-bin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622434 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-os-release\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622553 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-bin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-cnibin\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622656 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622710 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-cnibin\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622739 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-multus\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622763 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4kv8\" (UniqueName: \"kubernetes.io/projected/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-kube-api-access-f4kv8\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622812 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-cni-multus\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622777 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-binary-copy\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622940 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqlqr\" (UniqueName: \"kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cni-binary-copy\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.622954 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623435 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-binary-copy\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-hostroot\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-daemon-config\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623519 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-multus-certs\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-hostroot\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fab2e099-b326-498e-bd92-96260472132c-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-multus-certs\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623785 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-netns\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623799 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-rootfs\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623812 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8bxc\" (UniqueName: \"kubernetes.io/projected/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-kube-api-access-s8bxc\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cni-binary-copy\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623873 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-system-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623887 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-socket-dir-parent\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623901 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-proxy-tls\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623915 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-conf-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623916 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-rootfs\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623929 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623945 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623951 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623959 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623978 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.623992 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624006 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb0a03c5-46d9-4815-ab49-a93336faeff2-hosts-file\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624035 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5q2w\" (UniqueName: \"kubernetes.io/projected/bb0a03c5-46d9-4815-ab49-a93336faeff2-kube-api-access-l5q2w\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624048 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624072 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-mcd-auth-proxy-config\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624086 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfzq4\" (UniqueName: \"kubernetes.io/projected/fab2e099-b326-498e-bd92-96260472132c-kube-api-access-xfzq4\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624113 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624127 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624148 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624174 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cnibin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-os-release\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624210 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-kubelet\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624250 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-var-lib-kubelet\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624474 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624553 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-system-cni-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624581 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-socket-dir-parent\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624641 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-host-run-netns\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624802 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-conf-dir\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624828 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624848 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.624872 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625227 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-multus-daemon-config\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625402 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fab2e099-b326-498e-bd92-96260472132c-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625438 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625460 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/bb0a03c5-46d9-4815-ab49-a93336faeff2-hosts-file\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625692 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-cnibin\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625756 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625796 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-os-release\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625831 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625868 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.625942 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-mcd-auth-proxy-config\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.637282 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-proxy-tls\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.637566 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.646567 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e66d85e3e1958f34b2d4f147c4d9d6728fe5a6d28c3b2752460cb5ceb12310\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:40Z\\\",\\\"message\\\":\\\"W1004 03:40:29.612852 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 03:40:29.613285 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759549229 cert, and key in /tmp/serving-cert-744945107/serving-signer.crt, /tmp/serving-cert-744945107/serving-signer.key\\\\nI1004 03:40:29.965438 1 observer_polling.go:159] Starting file observer\\\\nW1004 03:40:29.972250 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 03:40:29.972510 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:29.977271 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-744945107/tls.crt::/tmp/serving-cert-744945107/tls.key\\\\\\\"\\\\nF1004 03:40:40.211393 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.646895 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5q2w\" (UniqueName: \"kubernetes.io/projected/bb0a03c5-46d9-4815-ab49-a93336faeff2-kube-api-access-l5q2w\") pod \"node-resolver-95wxp\" (UID: \"bb0a03c5-46d9-4815-ab49-a93336faeff2\") " pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.647727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqlqr\" (UniqueName: \"kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr\") pod \"ovnkube-node-snjd6\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.648147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4kv8\" (UniqueName: \"kubernetes.io/projected/e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1-kube-api-access-f4kv8\") pod \"machine-config-daemon-9777l\" (UID: \"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\") " pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.653085 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfzq4\" (UniqueName: \"kubernetes.io/projected/fab2e099-b326-498e-bd92-96260472132c-kube-api-access-xfzq4\") pod \"multus-additional-cni-plugins-t24l4\" (UID: \"fab2e099-b326-498e-bd92-96260472132c\") " pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.656535 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8bxc\" (UniqueName: \"kubernetes.io/projected/7fa9344a-8c9c-4bc5-b91e-b6c1400d4025-kube-api-access-s8bxc\") pod \"multus-5c96v\" (UID: \"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\") " pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.657544 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.659055 4726 scope.go:117] "RemoveContainer" containerID="7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa" Oct 04 03:40:48 crc kubenswrapper[4726]: E1004 03:40:48.659196 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.661426 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.662585 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.662613 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.662623 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"20d5aff6a33fd6b42b3f49b139edbff1fd364195f02a73db9ca07d2b7ccf8666"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.666092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d20d0588c9822d6172019c8e9c5aa08a914862e30dbd7cf8d5048edb0418a53a"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.667137 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.667168 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"68e02cbca47c54e5fc4faab0c4c6f21517f0f5b6354b826e44ce1f79e3efbc1f"} Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.687852 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.705473 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.727581 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.733579 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-95wxp" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.742271 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:48 crc kubenswrapper[4726]: W1004 03:40:48.746366 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb0a03c5_46d9_4815_ab49_a93336faeff2.slice/crio-5ad0f81cd99f4c921933b5e9230c98495a058afe2972fab6760e0f4148e5166f WatchSource:0}: Error finding container 5ad0f81cd99f4c921933b5e9230c98495a058afe2972fab6760e0f4148e5166f: Status 404 returned error can't find the container with id 5ad0f81cd99f4c921933b5e9230c98495a058afe2972fab6760e0f4148e5166f Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.751226 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.760015 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5c96v" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.763646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.765905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t24l4" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.785805 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: W1004 03:40:48.799242 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa9344a_8c9c_4bc5_b91e_b6c1400d4025.slice/crio-2bec6091caa21bb299fb56c4f93a8625a1ba451ea83a7e38af29a011ae97a7e5 WatchSource:0}: Error finding container 2bec6091caa21bb299fb56c4f93a8625a1ba451ea83a7e38af29a011ae97a7e5: Status 404 returned error can't find the container with id 2bec6091caa21bb299fb56c4f93a8625a1ba451ea83a7e38af29a011ae97a7e5 Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.806375 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.855441 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.870014 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.888280 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.900651 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.910482 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.921949 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.939673 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.951345 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.965399 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.983202 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:48 crc kubenswrapper[4726]: I1004 03:40:48.998956 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:48Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.129722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.129930 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:40:51.129900321 +0000 UTC m=+25.304523524 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.130165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.130199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.130223 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.130249 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130377 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130394 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130409 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130426 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130448 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:51.130439585 +0000 UTC m=+25.305062798 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130394 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130477 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:51.130461086 +0000 UTC m=+25.305084389 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130474 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130529 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130549 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130491 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:51.130486227 +0000 UTC m=+25.305109440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.130658 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:51.13063309 +0000 UTC m=+25.305256333 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.501537 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.501660 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.501687 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.501788 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.501957 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:49 crc kubenswrapper[4726]: E1004 03:40:49.502201 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.670962 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013" exitCode=0 Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.671038 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.671118 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"cade49fd4bdcc0cebe8a3471eaa37abe150a63fcdb9330da59719a47e6209fed"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.672340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-95wxp" event={"ID":"bb0a03c5-46d9-4815-ab49-a93336faeff2","Type":"ContainerStarted","Data":"8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.672379 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-95wxp" event={"ID":"bb0a03c5-46d9-4815-ab49-a93336faeff2","Type":"ContainerStarted","Data":"5ad0f81cd99f4c921933b5e9230c98495a058afe2972fab6760e0f4148e5166f"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.673745 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerStarted","Data":"6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.673766 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerStarted","Data":"2bec6091caa21bb299fb56c4f93a8625a1ba451ea83a7e38af29a011ae97a7e5"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.675160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.675195 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.675208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"637122a561870748f6386ff83702353aeb1ed52e45d1e0b49b936cd82c122902"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.676669 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35" exitCode=0 Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.676697 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.676710 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerStarted","Data":"56bc5c85e16404f31eb1304d965afec84929d350cdc235ca5bff80500caac090"} Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.696706 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.712051 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.725872 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.740044 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.758726 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.782506 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.801148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.813480 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.823430 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.837448 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.857075 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.869998 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.880772 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.891830 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.902401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.919376 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.932067 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.958040 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.970466 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:49 crc kubenswrapper[4726]: I1004 03:40:49.985758 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.000656 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:49Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.013593 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.026031 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.041316 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.682233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.684786 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerStarted","Data":"ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.687956 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.689158 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904" exitCode=1 Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.689205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.689231 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.689247 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.689259 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb"} Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.693016 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-b9rkg"] Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.694091 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.699125 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.699441 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.699609 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.699659 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.746554 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.775303 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.793166 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.805673 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.819813 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.838928 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.844540 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32c5bb99-f0d4-4259-8604-5b8664e20a5d-serviceca\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.844578 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbkw9\" (UniqueName: \"kubernetes.io/projected/32c5bb99-f0d4-4259-8604-5b8664e20a5d-kube-api-access-qbkw9\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.844637 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32c5bb99-f0d4-4259-8604-5b8664e20a5d-host\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.850887 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.865233 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.876531 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.888489 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.900707 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.908832 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.927614 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.940732 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.946068 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32c5bb99-f0d4-4259-8604-5b8664e20a5d-serviceca\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.946144 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbkw9\" (UniqueName: \"kubernetes.io/projected/32c5bb99-f0d4-4259-8604-5b8664e20a5d-kube-api-access-qbkw9\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.946204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32c5bb99-f0d4-4259-8604-5b8664e20a5d-host\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.946291 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32c5bb99-f0d4-4259-8604-5b8664e20a5d-host\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.947591 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32c5bb99-f0d4-4259-8604-5b8664e20a5d-serviceca\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.951261 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.963864 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.967614 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbkw9\" (UniqueName: \"kubernetes.io/projected/32c5bb99-f0d4-4259-8604-5b8664e20a5d-kube-api-access-qbkw9\") pod \"node-ca-b9rkg\" (UID: \"32c5bb99-f0d4-4259-8604-5b8664e20a5d\") " pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.974598 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.986445 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:50 crc kubenswrapper[4726]: I1004 03:40:50.998922 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:50Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.009985 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.022150 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-b9rkg" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.024204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.038386 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.052929 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.067831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.081010 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.150014 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.150160 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150197 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:40:55.150171596 +0000 UTC m=+29.324794809 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.150239 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.150280 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150295 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150316 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150328 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150375 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.150301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150376 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:55.150361391 +0000 UTC m=+29.324984734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150567 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:55.150550096 +0000 UTC m=+29.325173309 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150428 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150623 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:55.150616798 +0000 UTC m=+29.325240131 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150445 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150650 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150661 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.150691 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:40:55.15068573 +0000 UTC m=+29.325308943 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.501670 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.501746 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.501828 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.501920 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.502141 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:51 crc kubenswrapper[4726]: E1004 03:40:51.502853 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.698326 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b9rkg" event={"ID":"32c5bb99-f0d4-4259-8604-5b8664e20a5d","Type":"ContainerStarted","Data":"a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a"} Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.698664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-b9rkg" event={"ID":"32c5bb99-f0d4-4259-8604-5b8664e20a5d","Type":"ContainerStarted","Data":"f3bb84561e971efc1e8e81694f00870836b41c01bf445993c8b19838c74f9397"} Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.701838 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9" exitCode=0 Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.701972 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9"} Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.705992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.707378 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3"} Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.707407 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea"} Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.720854 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.741253 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.751093 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.761917 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.774989 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.786288 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.797370 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.810148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.825364 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.840614 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.852713 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.866310 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.876890 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.886219 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.898644 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.917496 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.934414 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.947229 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.961829 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.975209 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:51 crc kubenswrapper[4726]: I1004 03:40:51.987142 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:51.999999 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:51Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.012159 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.027757 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.038015 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.054028 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.712361 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db" exitCode=0 Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.712410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db"} Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.725194 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.752508 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.772600 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.789585 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.828683 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.852570 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.875336 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.889135 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.898366 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.910124 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.921191 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.930468 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:52 crc kubenswrapper[4726]: I1004 03:40:52.942510 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:52Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.089299 4726 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.091182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.091221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.091233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.091319 4726 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.099382 4726 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.099625 4726 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.100682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.100782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.100843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.100915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.100976 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.123794 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.127697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.127756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.127773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.127799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.127816 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.141864 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.145630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.145840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.145929 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.146017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.146140 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.163531 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.168557 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.168665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.168730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.168873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.168931 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.183011 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.186840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.186867 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.186876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.186889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.186902 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.199300 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.199435 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.201350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.201390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.201403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.201420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.201433 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.304350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.304417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.304439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.304467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.304485 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.407792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.407838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.407852 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.407873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.407887 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.469937 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.475981 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.479125 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.491005 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.502032 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.502044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.502124 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.502238 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.502354 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.502422 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.510244 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.510286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.510298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.510317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.510331 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.511683 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.534051 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.553287 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.570737 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.586719 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.608521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.613768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.613825 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.613846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.613871 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.613888 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.631765 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.650684 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.671259 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.692563 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.707152 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.716168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.716196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.716207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.716223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.716234 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.718976 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.719698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.720127 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.723214 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44" exitCode=0 Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.723350 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.734024 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.758981 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.775694 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.788702 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.807149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.808047 4726 scope.go:117] "RemoveContainer" containerID="7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa" Oct 04 03:40:53 crc kubenswrapper[4726]: E1004 03:40:53.808264 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.808313 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.823289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.823509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.823618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.823719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.823820 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.824658 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.844007 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.869588 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.883893 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.898911 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.922226 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.925913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.925949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.925958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.925972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.925981 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:53Z","lastTransitionTime":"2025-10-04T03:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.939270 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.955389 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.972574 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:53 crc kubenswrapper[4726]: I1004 03:40:53.990994 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:53Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.005082 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.018871 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.028313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.028489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.028602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.028722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.028834 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.032954 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.043665 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.058176 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.072926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.103928 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.124931 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.131699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.131756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.131766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.131786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.131799 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.137864 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.159940 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.193886 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.214858 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.234403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.234440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.234448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.234465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.234474 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.237450 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.336593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.336632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.336645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.336661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.336673 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.439531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.439608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.439633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.439662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.439683 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.542721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.542784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.542801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.542823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.542843 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.652301 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.652360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.652380 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.652405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.652423 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.730682 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerStarted","Data":"cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.755313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.755377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.755398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.755424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.755443 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.763528 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.790555 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.809371 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.829755 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.847238 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.857909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.857949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.857971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.857989 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.858001 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.874223 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.892541 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.905165 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.916526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.929578 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.943711 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.955171 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.959865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.959909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.959921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.959939 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.959952 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:54Z","lastTransitionTime":"2025-10-04T03:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.966740 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:54 crc kubenswrapper[4726]: I1004 03:40:54.986208 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.062360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.062437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.062464 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.062494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.062516 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.165317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.165360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.165370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.165389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.165401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.199050 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.199172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.199204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.199231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.199254 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199389 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199423 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199464 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199486 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199444 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.199427785 +0000 UTC m=+37.374050998 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199659 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.19963547 +0000 UTC m=+37.374258723 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199683 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.199670591 +0000 UTC m=+37.374293854 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199743 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199786 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.199773084 +0000 UTC m=+37.374396337 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199861 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199879 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199892 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.199937 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.199926738 +0000 UTC m=+37.374550061 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.271306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.271365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.271385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.271411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.271430 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.374575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.374887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.375048 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.375323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.375492 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.478740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.478808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.478828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.478853 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.478870 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.502047 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.502062 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.502433 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.502182 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.502771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:55 crc kubenswrapper[4726]: E1004 03:40:55.502628 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.582204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.582279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.582298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.582325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.582344 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.685397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.685449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.685461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.685478 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.685490 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.737209 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6" exitCode=0 Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.737263 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.750966 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.777819 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.790967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.791010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.791022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.791041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.791053 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.801521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.819314 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.834596 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.851046 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.867884 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.883848 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.893219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.893245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.893254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.893268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.893278 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.897772 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.914920 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.930693 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.946295 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.960364 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.970518 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.996755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.996805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.996820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.996839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:55 crc kubenswrapper[4726]: I1004 03:40:55.996853 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:55Z","lastTransitionTime":"2025-10-04T03:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.098713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.098791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.098807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.098828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.098846 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.201436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.201480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.201491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.201509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.201522 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.304375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.304422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.304437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.304456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.304471 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.407337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.407401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.407419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.407445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.407463 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.509636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.509698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.509752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.509780 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.509802 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.520820 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.537450 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.553379 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.565909 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.581355 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.600703 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.614026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.614412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.614727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.614971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.615201 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.615737 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.640508 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.660022 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.674004 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.685396 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.711036 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.718460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.718499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.718507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.718522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.718532 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.728398 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.747905 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.751706 4726 generic.go:334] "Generic (PLEG): container finished" podID="fab2e099-b326-498e-bd92-96260472132c" containerID="5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6" exitCode=0 Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.752174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerDied","Data":"5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.761851 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.763682 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.764102 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.764389 4726 scope.go:117] "RemoveContainer" containerID="c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.770440 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.788378 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.792547 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.803756 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.818454 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.821220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.821434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.821571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.821696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.822130 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.832040 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.850278 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.865006 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.875364 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.886576 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.899322 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.911248 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.924019 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.925361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.925393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.925401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.925417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.925429 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:56Z","lastTransitionTime":"2025-10-04T03:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.936586 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.947401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.968066 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-acl-logging nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:56 crc kubenswrapper[4726]: I1004 03:40:56.990424 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.000648 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:56Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.011786 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.022869 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.029005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.029162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.029190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.029259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.029294 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.034759 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.046586 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.061081 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.073419 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.085658 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.099261 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.111701 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.129538 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.131718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.131771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.131783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.131800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.131829 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.144148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.234159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.234431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.234520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.234690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.234794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.338157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.338291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.338370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.338404 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.338480 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.442782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.442830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.442842 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.442859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.442872 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.501926 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:57 crc kubenswrapper[4726]: E1004 03:40:57.502065 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.502494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:57 crc kubenswrapper[4726]: E1004 03:40:57.502572 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.502620 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:57 crc kubenswrapper[4726]: E1004 03:40:57.502672 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.546501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.546561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.546581 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.546608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.546631 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.652434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.652475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.652486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.652502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.652514 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.755212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.755250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.755265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.755283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.755298 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.779884 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.780912 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.781283 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.781352 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.786451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" event={"ID":"fab2e099-b326-498e-bd92-96260472132c","Type":"ContainerStarted","Data":"5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.796477 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.808666 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.819928 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.840061 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.857403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.857451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.857462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.857483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.857496 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.862789 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.881949 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.900456 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.916258 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.931737 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.947826 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.960035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.960313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.960333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.960390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.960405 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:57Z","lastTransitionTime":"2025-10-04T03:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.963840 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.979555 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:57 crc kubenswrapper[4726]: I1004 03:40:57.997219 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:57Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.012135 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.024679 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.039136 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.055939 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.062825 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.062856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.062868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.062884 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.062895 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.067937 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.085722 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.096339 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.113574 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.123915 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.143188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.154274 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164084 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.164766 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.177492 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.189319 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.201948 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.213170 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:40:58Z is after 2025-08-24T17:21:41Z" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.267750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.267807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.267824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.267849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.267866 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.370938 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.371009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.371030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.371056 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.371075 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.473417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.473456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.473467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.473483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.473492 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.576755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.576821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.576831 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.576852 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.576864 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.680016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.680057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.680066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.680081 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.680092 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.782769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.782846 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.782866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.782896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.782915 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.885654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.885716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.885739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.885767 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.885790 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.988362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.988409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.988420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.988437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:58 crc kubenswrapper[4726]: I1004 03:40:58.988448 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:58Z","lastTransitionTime":"2025-10-04T03:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.090498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.090531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.090539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.090551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.090559 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.193338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.193424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.193459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.193496 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.193519 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.296138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.296190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.296201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.296219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.296230 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.399459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.399530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.399559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.399586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.399604 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.501721 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.501787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.501947 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:40:59 crc kubenswrapper[4726]: E1004 03:40:59.502095 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:40:59 crc kubenswrapper[4726]: E1004 03:40:59.502246 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:40:59 crc kubenswrapper[4726]: E1004 03:40:59.502362 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.502412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.502512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.502568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.502590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.502643 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.606233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.606328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.606350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.606381 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.606401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.709885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.709929 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.709942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.709960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.709973 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.812004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.812042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.812053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.812069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.812081 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.915091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.915199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.915217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.915624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:40:59 crc kubenswrapper[4726]: I1004 03:40:59.915678 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:40:59Z","lastTransitionTime":"2025-10-04T03:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.018419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.018479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.018501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.018533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.018553 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.121871 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.122343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.122516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.122691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.122858 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.225848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.225911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.225931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.225959 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.225981 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.329499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.329561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.329584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.329612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.329633 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.432497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.432567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.432584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.432608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.432630 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.536323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.536526 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.536561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.536638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.536662 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.640099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.640201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.640218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.640272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.640290 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.743582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.743643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.743661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.743688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.743705 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.846775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.846838 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.846856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.846887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.846904 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.950424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.950493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.950510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.950564 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:00 crc kubenswrapper[4726]: I1004 03:41:00.950582 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:00Z","lastTransitionTime":"2025-10-04T03:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.053730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.053794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.053812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.053837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.053854 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.156727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.157034 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.157271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.157498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.157684 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.261362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.261442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.261484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.261513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.261530 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.328462 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd"] Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.329081 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.336850 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.336898 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.355239 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.363855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.363926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.363942 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.363966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.364007 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.368001 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.368082 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.368144 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.368173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr6f9\" (UniqueName: \"kubernetes.io/projected/13f6cac4-0dd4-466d-820f-0822d4e0128f-kube-api-access-xr6f9\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.385645 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.410232 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.427255 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.442742 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.453990 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.466074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.466149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.466165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.466182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.466194 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.468934 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.469127 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.469191 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.469222 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr6f9\" (UniqueName: \"kubernetes.io/projected/13f6cac4-0dd4-466d-820f-0822d4e0128f-kube-api-access-xr6f9\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.469253 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.469950 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.470069 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/13f6cac4-0dd4-466d-820f-0822d4e0128f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.475926 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/13f6cac4-0dd4-466d-820f-0822d4e0128f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.486281 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.494634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr6f9\" (UniqueName: \"kubernetes.io/projected/13f6cac4-0dd4-466d-820f-0822d4e0128f-kube-api-access-xr6f9\") pod \"ovnkube-control-plane-749d76644c-xvgfd\" (UID: \"13f6cac4-0dd4-466d-820f-0822d4e0128f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.501993 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.502020 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.502071 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:01 crc kubenswrapper[4726]: E1004 03:41:01.502416 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.502478 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: E1004 03:41:01.502622 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:01 crc kubenswrapper[4726]: E1004 03:41:01.502727 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.521192 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.537637 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.557871 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.571713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.571747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.571756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.571772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.571782 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.574909 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.585783 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.597867 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.644172 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" Oct 04 03:41:01 crc kubenswrapper[4726]: W1004 03:41:01.663928 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13f6cac4_0dd4_466d_820f_0822d4e0128f.slice/crio-3210ada7eaa24055eff9d550e0965520c41627e0170d45790ce8f46c5b71db76 WatchSource:0}: Error finding container 3210ada7eaa24055eff9d550e0965520c41627e0170d45790ce8f46c5b71db76: Status 404 returned error can't find the container with id 3210ada7eaa24055eff9d550e0965520c41627e0170d45790ce8f46c5b71db76 Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.674079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.674158 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.674177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.674200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.674216 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.776443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.776502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.776520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.776545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.776563 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.807311 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" event={"ID":"13f6cac4-0dd4-466d-820f-0822d4e0128f","Type":"ContainerStarted","Data":"3210ada7eaa24055eff9d550e0965520c41627e0170d45790ce8f46c5b71db76"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.809813 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/0.log" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.812871 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.813461 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd" exitCode=1 Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.813492 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.814145 4726 scope.go:117] "RemoveContainer" containerID="8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.833287 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.849298 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.864605 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.878509 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.879817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.879865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.879883 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.879906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.879921 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.894533 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.907387 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.931050 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:00Z\\\",\\\"message\\\":\\\"0.087699 5985 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.087864 5985 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088290 5985 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088483 5985 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088688 5985 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088782 5985 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.089252 5985 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:00.089372 5985 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:41:00.089411 5985 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:00.089433 5985 factory.go:656] Stopping watch factory\\\\nI1004 03:41:00.089451 5985 ovnkube.go:599] Stopped ovnkube\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.951069 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.963303 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.982662 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.983433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.983493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.983512 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.983536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:01 crc kubenswrapper[4726]: I1004 03:41:01.983558 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:01Z","lastTransitionTime":"2025-10-04T03:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.001464 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:01Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.018892 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.035831 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.056130 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.074185 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.087026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.087134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.087163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.087196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.087220 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.189426 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.189461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.189472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.189488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.189509 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.292751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.292807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.292821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.292847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.292858 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.395293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.395317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.395328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.395339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.395348 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.497367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.497406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.497414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.497431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.497441 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.599922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.600008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.600028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.600049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.600065 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.703174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.703238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.703256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.703282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.703300 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.797991 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9qn78"] Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.798669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:02 crc kubenswrapper[4726]: E1004 03:41:02.798763 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.806046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.806267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.806363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.806417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.806442 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.820077 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" event={"ID":"13f6cac4-0dd4-466d-820f-0822d4e0128f","Type":"ContainerStarted","Data":"02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.820145 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" event={"ID":"13f6cac4-0dd4-466d-820f-0822d4e0128f","Type":"ContainerStarted","Data":"9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.823441 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/0.log" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.826733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.827756 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.828252 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.829841 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:00Z\\\",\\\"message\\\":\\\"0.087699 5985 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.087864 5985 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088290 5985 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088483 5985 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088688 5985 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088782 5985 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.089252 5985 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:00.089372 5985 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:41:00.089411 5985 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:00.089433 5985 factory.go:656] Stopping watch factory\\\\nI1004 03:41:00.089451 5985 ovnkube.go:599] Stopped ovnkube\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.846257 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.858759 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.878470 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.885313 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22dkh\" (UniqueName: \"kubernetes.io/projected/77b969b7-0698-4589-a6cf-c08cc779ffe2-kube-api-access-22dkh\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.885585 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.895313 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908394 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908855 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.908897 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:02Z","lastTransitionTime":"2025-10-04T03:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.923381 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.939889 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.959531 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.978634 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.987044 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.987179 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22dkh\" (UniqueName: \"kubernetes.io/projected/77b969b7-0698-4589-a6cf-c08cc779ffe2-kube-api-access-22dkh\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:02 crc kubenswrapper[4726]: E1004 03:41:02.987260 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:02 crc kubenswrapper[4726]: E1004 03:41:02.987384 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:03.487355549 +0000 UTC m=+37.661978802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:02 crc kubenswrapper[4726]: I1004 03:41:02.995325 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:02Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.011692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.011781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.011801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.011824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.011874 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.019344 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.019781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22dkh\" (UniqueName: \"kubernetes.io/projected/77b969b7-0698-4589-a6cf-c08cc779ffe2-kube-api-access-22dkh\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.039991 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.058459 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.073729 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.090888 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.114667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.114715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.114734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.114757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.114774 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.120432 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:00Z\\\",\\\"message\\\":\\\"0.087699 5985 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.087864 5985 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088290 5985 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088483 5985 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088688 5985 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088782 5985 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.089252 5985 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:00.089372 5985 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:41:00.089411 5985 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:00.089433 5985 factory.go:656] Stopping watch factory\\\\nI1004 03:41:00.089451 5985 ovnkube.go:599] Stopped ovnkube\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.142257 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.154406 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.173756 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.187988 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.208918 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.217460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.217489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.217497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.217509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.217518 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.228727 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.246972 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.261231 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.280604 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.289378 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.289482 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.289513 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.289555 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.289579 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289617 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:41:19.289578441 +0000 UTC m=+53.464201664 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289700 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289715 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289743 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:19.289733065 +0000 UTC m=+53.464356288 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289766 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:19.289750536 +0000 UTC m=+53.464373869 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289809 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289857 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289871 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289887 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289946 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.289973 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.290021 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:19.289989862 +0000 UTC m=+53.464613115 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.290062 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:19.290043423 +0000 UTC m=+53.464666756 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.299830 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.318899 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.320091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.320327 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.320398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.320425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.320441 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.338145 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.358403 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.374544 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.391781 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.423032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.423118 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.423137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.423178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.423195 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.477409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.477471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.477489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.477514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.477533 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.491928 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.492186 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.492283 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:04.492258735 +0000 UTC m=+38.666881978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.499763 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.501739 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.501795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.501795 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.501931 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.502089 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.502294 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.505977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.506043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.506067 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.506093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.506155 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.522300 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.527349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.527403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.527419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.527443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.527460 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.542312 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.545875 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.545968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.545987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.546028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.546045 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.562021 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.565859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.565895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.565907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.565922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.565932 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.579584 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.579792 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.581737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.581777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.581789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.581804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.581816 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.684748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.684812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.684829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.684857 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.684875 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.788072 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.788147 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.788163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.788184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.788199 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.833000 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/1.log" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.833782 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/0.log" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.835276 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.835763 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889" exitCode=1 Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.836661 4726 scope.go:117] "RemoveContainer" containerID="94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889" Oct 04 03:41:03 crc kubenswrapper[4726]: E1004 03:41:03.836776 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.836818 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.836841 4726 scope.go:117] "RemoveContainer" containerID="8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.848494 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.863515 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.872706 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.888924 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a84ec2148115810bd498f2e8d288b913c4ff1ca721f6903bf745b2e5c1eeabd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:00Z\\\",\\\"message\\\":\\\"0.087699 5985 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.087864 5985 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088290 5985 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088483 5985 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088688 5985 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.088782 5985 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:00.089252 5985 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:00.089372 5985 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:00.089384 5985 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 03:41:00.089411 5985 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:00.089433 5985 factory.go:656] Stopping watch factory\\\\nI1004 03:41:00.089451 5985 ovnkube.go:599] Stopped ovnkube\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.890284 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.890329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.890345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.890367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.890383 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.905782 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.920662 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.938490 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.956023 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.973743 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.987209 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.992550 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.992579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.992588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.992619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.992631 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:03Z","lastTransitionTime":"2025-10-04T03:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:03 crc kubenswrapper[4726]: I1004 03:41:03.997121 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:03Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.012937 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.032602 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.047145 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.062571 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.082336 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.094323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.094347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.094355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.094368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.094376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.196410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.196453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.196465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.196479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.196490 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.298790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.298820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.298828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.298839 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.298847 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.402391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.402490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.402517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.402549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.402570 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.501911 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:04 crc kubenswrapper[4726]: E1004 03:41:04.502202 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.502205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:04 crc kubenswrapper[4726]: E1004 03:41:04.502398 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:04 crc kubenswrapper[4726]: E1004 03:41:04.502496 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:06.50246765 +0000 UTC m=+40.677090903 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.504316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.504359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.504369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.504385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.504398 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.607392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.607437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.607451 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.607470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.607484 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.710918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.710980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.710997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.711022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.711039 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.813469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.813534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.813551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.813576 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.813594 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.848192 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/1.log" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.852246 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.854690 4726 scope.go:117] "RemoveContainer" containerID="94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889" Oct 04 03:41:04 crc kubenswrapper[4726]: E1004 03:41:04.854960 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.880481 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.903972 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.916985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.917052 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.917075 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.917135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.917161 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:04Z","lastTransitionTime":"2025-10-04T03:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.923891 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.943341 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.961221 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:04 crc kubenswrapper[4726]: I1004 03:41:04.977089 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:04Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.004638 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.020220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.020288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.020305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.020330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.020347 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.023411 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.037088 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.055732 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.069188 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.081203 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.090987 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.103184 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.115864 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.122401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.122450 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.122465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.122486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.122500 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.128191 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:05Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.224786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.224829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.224843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.224862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.224877 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.327126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.327189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.327204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.327224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.327238 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.430472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.430533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.430551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.430573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.430590 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.502001 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.502053 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.502053 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:05 crc kubenswrapper[4726]: E1004 03:41:05.502221 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:05 crc kubenswrapper[4726]: E1004 03:41:05.502349 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:05 crc kubenswrapper[4726]: E1004 03:41:05.502470 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.533017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.533079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.533126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.533152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.533170 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.635889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.635951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.635969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.635995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.636014 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.739441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.739505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.739528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.739558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.739582 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.842772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.842868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.842890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.842924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.842945 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.945607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.945676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.945699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.945727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:05 crc kubenswrapper[4726]: I1004 03:41:05.945749 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:05Z","lastTransitionTime":"2025-10-04T03:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.049248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.049308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.049334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.049365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.049387 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.151543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.151589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.151604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.151624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.151637 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.254743 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.254830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.254848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.254874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.254893 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.357850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.357974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.357998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.358031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.358052 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.462200 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.462332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.462411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.462446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.462546 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.502050 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:06 crc kubenswrapper[4726]: E1004 03:41:06.502315 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.521175 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.525904 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:06 crc kubenswrapper[4726]: E1004 03:41:06.526170 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:06 crc kubenswrapper[4726]: E1004 03:41:06.526261 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:10.526233997 +0000 UTC m=+44.700857240 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.539758 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.556585 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.565521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.565833 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.565970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.566156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.566305 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.587590 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.612348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.641182 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.660491 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.668980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.669026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.669038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.669057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.669069 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.679528 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.699342 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.718431 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.740253 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.764419 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.771184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.771345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.771431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.771577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.771886 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.780775 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.794581 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.807085 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.821625 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:06Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.874948 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.875008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.875023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.875047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.875065 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.978472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.978556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.978583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.978614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:06 crc kubenswrapper[4726]: I1004 03:41:06.978636 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:06Z","lastTransitionTime":"2025-10-04T03:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.081822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.082010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.082037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.082068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.082091 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.185193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.185504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.185699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.185996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.186246 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.289606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.289665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.289683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.289706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.289726 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.393174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.393234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.393252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.393276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.393293 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.496589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.496636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.496646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.496662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.496674 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.502252 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:07 crc kubenswrapper[4726]: E1004 03:41:07.502407 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.502494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.502495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:07 crc kubenswrapper[4726]: E1004 03:41:07.502970 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:07 crc kubenswrapper[4726]: E1004 03:41:07.503062 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.503619 4726 scope.go:117] "RemoveContainer" containerID="7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.600179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.600250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.600276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.600305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.600323 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.703315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.703377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.703395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.703422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.703442 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.807040 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.807534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.807551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.807579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.807597 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.869793 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.875101 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.875740 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.905008 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.910794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.910860 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.910880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.910910 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.910930 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:07Z","lastTransitionTime":"2025-10-04T03:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.925572 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.946083 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.966254 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:07 crc kubenswrapper[4726]: I1004 03:41:07.982199 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:07.999879 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:07Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.014265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.014309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.014319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.014338 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.014352 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.025552 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.046260 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.061862 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.080142 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.096570 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.111092 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.117844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.117915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.117944 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.117975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.118002 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.131793 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.152378 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.199031 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.220995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.221071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.221095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.221153 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.221174 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.224028 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:08Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.324091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.324383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.324495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.324588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.324668 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.427424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.427680 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.427762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.427840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.427921 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.502404 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:08 crc kubenswrapper[4726]: E1004 03:41:08.502778 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.530383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.530559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.530579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.530597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.530609 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.633329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.633386 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.633403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.633424 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.633442 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.735955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.736357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.736509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.736660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.736806 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.839894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.839971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.839997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.840027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.840051 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.942546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.942872 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.943010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.943198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:08 crc kubenswrapper[4726]: I1004 03:41:08.943349 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:08Z","lastTransitionTime":"2025-10-04T03:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.046354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.046405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.046417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.046438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.046451 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.149096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.149170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.149183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.149202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.149213 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.253016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.253079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.253102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.253169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.253191 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.356980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.357049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.357071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.357102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.357194 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.460983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.461170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.461191 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.461220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.461241 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.501646 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.501756 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:09 crc kubenswrapper[4726]: E1004 03:41:09.501868 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:09 crc kubenswrapper[4726]: E1004 03:41:09.502090 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.501694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:09 crc kubenswrapper[4726]: E1004 03:41:09.502520 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.564681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.564740 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.564760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.564786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.564809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.667891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.667969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.667988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.668018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.668037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.771673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.771803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.771831 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.771869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.771890 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.875957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.876002 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.876013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.876028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.876038 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.979411 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.979462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.979474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.979495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:09 crc kubenswrapper[4726]: I1004 03:41:09.979507 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:09Z","lastTransitionTime":"2025-10-04T03:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.082487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.082551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.082571 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.082596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.082617 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.186190 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.186263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.186333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.186370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.186393 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.292596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.292664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.292684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.292711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.292732 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.395996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.396065 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.396087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.396264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.396298 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.498823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.498881 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.498898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.498921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.498938 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.502600 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:10 crc kubenswrapper[4726]: E1004 03:41:10.502819 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.569522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:10 crc kubenswrapper[4726]: E1004 03:41:10.569780 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:10 crc kubenswrapper[4726]: E1004 03:41:10.570128 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:18.570078612 +0000 UTC m=+52.744701855 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.602082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.602169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.602192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.602215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.602232 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.705199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.705554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.705797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.706032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.706302 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.809563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.809638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.809660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.809692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.809713 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.913355 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.913449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.913469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.913941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:10 crc kubenswrapper[4726]: I1004 03:41:10.914020 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:10Z","lastTransitionTime":"2025-10-04T03:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.017215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.017255 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.017266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.017288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.017301 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.120806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.120859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.120877 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.120893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.120906 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.224705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.224775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.224793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.224837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.224860 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.329037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.329095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.329149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.329175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.329192 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.431992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.432091 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.432164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.432202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.432225 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.502352 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.502473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:11 crc kubenswrapper[4726]: E1004 03:41:11.502579 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.502406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:11 crc kubenswrapper[4726]: E1004 03:41:11.502702 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:11 crc kubenswrapper[4726]: E1004 03:41:11.502796 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.535649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.535726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.535751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.535782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.535807 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.638906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.638962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.638980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.639003 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.639022 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.742412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.742482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.742503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.742532 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.742554 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.845148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.845223 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.845246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.845277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.845365 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.948051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.948139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.948156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.948180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:11 crc kubenswrapper[4726]: I1004 03:41:11.948197 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:11Z","lastTransitionTime":"2025-10-04T03:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.051669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.051709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.051719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.051734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.051744 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.154578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.154614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.154624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.154639 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.154650 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.257050 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.257164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.257184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.257208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.257227 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.360267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.360612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.360787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.360972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.361171 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.464655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.464720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.464732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.464766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.464779 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.502481 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:12 crc kubenswrapper[4726]: E1004 03:41:12.502702 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.567393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.567471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.567490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.567515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.567531 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.670787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.670865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.670890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.670919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.670940 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.773573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.773633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.773652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.773675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.773694 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.876674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.876974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.877189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.877351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.877534 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.980468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.980518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.980536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.980558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:12 crc kubenswrapper[4726]: I1004 03:41:12.980577 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:12Z","lastTransitionTime":"2025-10-04T03:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.083178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.083212 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.083222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.083233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.083242 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.185967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.186057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.186080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.186136 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.186152 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.289474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.289539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.289559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.289589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.289610 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.393228 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.393275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.393286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.393303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.393315 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.496348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.496398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.496409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.496429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.496441 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.501890 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.501890 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.502095 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.502178 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.501918 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.502362 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.600156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.600247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.600258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.600276 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.600290 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.704224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.704293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.704310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.704335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.704351 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.741777 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.741841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.741859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.741891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.741912 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.759267 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:13Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.764485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.764551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.764573 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.764605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.764627 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.785416 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:13Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.791242 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.791601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.791815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.792050 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.792251 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.813139 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:13Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.818261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.818312 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.818326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.818347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.818357 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.838144 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:13Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.842943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.842994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.843010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.843030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.843043 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.863993 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:13Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:13 crc kubenswrapper[4726]: E1004 03:41:13.864161 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.866210 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.866262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.866273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.866293 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.866307 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.969271 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.969344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.969368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.969402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:13 crc kubenswrapper[4726]: I1004 03:41:13.969426 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:13Z","lastTransitionTime":"2025-10-04T03:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.072759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.072801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.072812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.072829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.072840 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.176352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.176403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.176420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.176445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.176462 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.279926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.279983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.280000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.280022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.280041 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.383955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.384398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.384566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.384756 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.384951 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.488354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.488406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.488422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.488445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.488464 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.502188 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:14 crc kubenswrapper[4726]: E1004 03:41:14.502446 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.591577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.591615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.591625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.591640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.591651 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.694204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.694446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.694566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.694650 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.694738 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.797230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.797301 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.797324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.797351 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.797368 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.900594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.900660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.900684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.900710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:14 crc kubenswrapper[4726]: I1004 03:41:14.900732 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:14Z","lastTransitionTime":"2025-10-04T03:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.003583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.003640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.003666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.003727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.003752 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.106702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.106786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.106818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.106850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.106871 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.209682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.209744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.209784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.209821 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.209834 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.312843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.312892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.312906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.312924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.312937 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.415605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.415690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.415714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.415746 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.415769 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.502038 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:15 crc kubenswrapper[4726]: E1004 03:41:15.502239 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.502372 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:15 crc kubenswrapper[4726]: E1004 03:41:15.502506 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.502579 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:15 crc kubenswrapper[4726]: E1004 03:41:15.502632 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.519565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.519722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.519739 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.519759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.519777 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.623375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.623463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.623486 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.623520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.623544 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.727600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.727673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.727701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.727771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.727827 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.830556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.830603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.830615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.830636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.830648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.934144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.934181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.934192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.934209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:15 crc kubenswrapper[4726]: I1004 03:41:15.934222 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:15Z","lastTransitionTime":"2025-10-04T03:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.036903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.036974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.036999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.037033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.037094 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.141011 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.141084 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.141139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.141216 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.141281 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.244721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.244805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.244826 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.244863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.244886 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.348428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.349688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.349859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.350013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.350202 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.453341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.453393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.453408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.453431 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.453442 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.502976 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:16 crc kubenswrapper[4726]: E1004 03:41:16.503157 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.533443 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.557175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.557253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.557273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.557309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.557329 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.560699 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.584054 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.607889 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.625717 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.648926 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.660529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.660595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.660605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.660630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.660641 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.688090 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.717177 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.737401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.754863 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.763737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.763783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.763803 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.763829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.763849 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.771589 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.790082 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.808713 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.828778 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.849280 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.867131 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.867215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.867241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.867273 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.867295 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.871514 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:16Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.969974 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.970045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.970069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.970099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:16 crc kubenswrapper[4726]: I1004 03:41:16.970158 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:16Z","lastTransitionTime":"2025-10-04T03:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.072570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.072634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.072669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.072707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.072731 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.175337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.175406 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.175428 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.175457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.175480 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.278922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.278981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.278997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.279020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.279037 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.382088 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.382219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.382248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.382280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.382305 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.485514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.485900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.486169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.486423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.486637 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.500398 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.501660 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:17 crc kubenswrapper[4726]: E1004 03:41:17.502019 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.501727 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.501693 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:17 crc kubenswrapper[4726]: E1004 03:41:17.502825 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:17 crc kubenswrapper[4726]: E1004 03:41:17.502617 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.514388 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.523941 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.541895 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.558183 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.577296 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.590138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.590421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.590559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.590695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.590855 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.597271 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.615053 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.632941 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.653954 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.677704 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.694968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.695058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.695077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.695101 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.695154 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.700562 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.718582 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.735348 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.748402 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.770098 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.787445 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.798062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.798143 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.798155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.798227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.798261 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.802970 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:17Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.900800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.901025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.901176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.901262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:17 crc kubenswrapper[4726]: I1004 03:41:17.901341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:17Z","lastTransitionTime":"2025-10-04T03:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.004181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.004251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.004272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.004299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.004317 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.107234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.107298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.107317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.107342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.107359 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.210906 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.210951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.210967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.210986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.211000 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.313649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.313712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.313730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.313751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.313769 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.417782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.417865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.417889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.418397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.418685 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.502534 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:18 crc kubenswrapper[4726]: E1004 03:41:18.502760 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.520961 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.521019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.521041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.521069 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.521094 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.624447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.624511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.624536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.624579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.624604 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.664297 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:18 crc kubenswrapper[4726]: E1004 03:41:18.664550 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:18 crc kubenswrapper[4726]: E1004 03:41:18.664664 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:34.664629526 +0000 UTC m=+68.839252779 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.727409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.727474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.727492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.727523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.727548 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.829854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.830183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.830308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.830427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.830537 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.933417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.933469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.933483 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.933501 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:18 crc kubenswrapper[4726]: I1004 03:41:18.933515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:18Z","lastTransitionTime":"2025-10-04T03:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.036258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.036329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.036352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.036382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.036405 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.139364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.139403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.139416 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.139433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.139445 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.242683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.242753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.242774 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.242802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.242822 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.347004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.347071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.347090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.347146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.347165 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.371418 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.371670 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:41:51.37159943 +0000 UTC m=+85.546222683 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.371955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.372184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372246 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372562 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372657 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.372519 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.372784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372383 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372970 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372997 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.373020 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.372836 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:51.37276506 +0000 UTC m=+85.547388313 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.373143 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:51.373080269 +0000 UTC m=+85.547703522 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.373184 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:51.373164021 +0000 UTC m=+85.547787424 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.373353 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.373460 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:41:51.373436598 +0000 UTC m=+85.548059841 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.451485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.451558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.451580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.451611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.451634 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.502335 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.502374 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.502460 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.503054 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.503172 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:19 crc kubenswrapper[4726]: E1004 03:41:19.503261 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.503596 4726 scope.go:117] "RemoveContainer" containerID="94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.554368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.554420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.554432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.554447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.554459 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.657703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.657753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.657772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.657796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.657814 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.759986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.760027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.760038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.760054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.760068 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.863540 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.863608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.863629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.863655 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.863675 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.920867 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/1.log" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.923399 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.924313 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093"} Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.924827 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.955868 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:19Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.967630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.967684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.967696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.967713 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:19 crc kubenswrapper[4726]: I1004 03:41:19.967727 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:19Z","lastTransitionTime":"2025-10-04T03:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.010429 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.027653 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.045736 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.061801 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.070415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.070472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.070485 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.070513 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.070529 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.080401 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.093534 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.111154 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.123084 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.137768 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.156457 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.169666 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.174138 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.174222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.174241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.174272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.174291 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.185700 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.205204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.221274 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.236028 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.249160 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.276882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.276925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.276937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.276952 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.276962 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.379814 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.379861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.379874 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.379893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.379904 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.482901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.482943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.482957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.482976 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.482989 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.502593 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:20 crc kubenswrapper[4726]: E1004 03:41:20.502800 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.585548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.585586 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.585595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.585608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.585618 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.687876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.687931 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.687941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.687956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.687965 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.791580 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.791651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.791669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.792066 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.792138 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.894589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.894666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.894683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.894710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.894728 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.928701 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/2.log" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.929522 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/1.log" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.932624 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.933466 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" exitCode=1 Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.933541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093"} Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.933602 4726 scope.go:117] "RemoveContainer" containerID="94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.934268 4726 scope.go:117] "RemoveContainer" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" Oct 04 03:41:20 crc kubenswrapper[4726]: E1004 03:41:20.934437 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.956750 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.976450 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.989962 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:20Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.997132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.997165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.997177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.997192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:20 crc kubenswrapper[4726]: I1004 03:41:20.997204 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:20Z","lastTransitionTime":"2025-10-04T03:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.004360 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.018229 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.030794 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.042617 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.060789 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94461c8da395e365aa949c5aed4bb32d5155961f1e97b88413fa4796b2ae2889\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:03Z\\\",\\\"message\\\":\\\"controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 03:41:02.873081 6182 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 03:41:02.873144 6182 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 03:41:02.873170 6182 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 03:41:02.873197 6182 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:02.873205 6182 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 03:41:02.873238 6182 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1004 03:41:02.873277 6182 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 03:41:02.873331 6182 factory.go:656] Stopping watch factory\\\\nI1004 03:41:02.873353 6182 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:02.873384 6182 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 03:41:02.873403 6182 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 03:41:02.873415 6182 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 03:41:02.873427 6182 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:02.873439 6182 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 03:41:02.873456 6182 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.078223 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.092704 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.099836 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.099870 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.099878 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.099891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.099902 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.111544 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.124175 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.141976 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.157340 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.168843 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.181851 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.194133 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.201827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.201891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.201911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.201937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.201955 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.304780 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.304845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.304869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.304896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.304916 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.409562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.409630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.409647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.409673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.409703 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.502297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.502384 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.502297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:21 crc kubenswrapper[4726]: E1004 03:41:21.502497 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:21 crc kubenswrapper[4726]: E1004 03:41:21.502664 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:21 crc kubenswrapper[4726]: E1004 03:41:21.502749 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.512737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.512783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.512796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.512816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.512832 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.616215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.616286 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.616305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.616335 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.616358 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.718757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.718830 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.718850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.718873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.718892 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.822342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.822412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.822435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.822463 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.822487 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.925572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.925627 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.925643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.925666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.925682 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:21Z","lastTransitionTime":"2025-10-04T03:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.940198 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/2.log" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.944723 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.947171 4726 scope.go:117] "RemoveContainer" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" Oct 04 03:41:21 crc kubenswrapper[4726]: E1004 03:41:21.947456 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:21 crc kubenswrapper[4726]: I1004 03:41:21.979451 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:21Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.004425 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.020972 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.027798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.027854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.027878 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.027907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.027928 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.041200 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.057689 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.072923 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.091559 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.111138 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.128845 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.130605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.130664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.130685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.130738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.130758 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.150096 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.170571 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.194261 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.210417 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.221262 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.231662 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.233308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.233343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.233353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.233366 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.233376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.241845 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.251736 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:22Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.336280 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.336353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.336371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.336397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.336417 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.438952 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.439035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.439054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.439082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.439144 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.502025 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:22 crc kubenswrapper[4726]: E1004 03:41:22.502323 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.542173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.542258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.542288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.542317 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.542337 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.645157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.645218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.645234 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.645259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.645278 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.747506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.747569 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.747585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.747606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.747623 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.850331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.850410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.850436 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.850468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.850492 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.953435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.953503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.953521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.953546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:22 crc kubenswrapper[4726]: I1004 03:41:22.953562 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:22Z","lastTransitionTime":"2025-10-04T03:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.056268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.056331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.056352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.056377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.056395 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.160403 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.160471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.160493 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.160523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.160545 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.263100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.263205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.263229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.263260 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.263281 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.366350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.366399 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.366409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.366427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.366483 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.468452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.468521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.468536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.468556 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.468571 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.502494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.502561 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:23 crc kubenswrapper[4726]: E1004 03:41:23.502627 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.502494 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:23 crc kubenswrapper[4726]: E1004 03:41:23.502699 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:23 crc kubenswrapper[4726]: E1004 03:41:23.502739 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.570460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.570503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.570518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.570534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.570543 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.672973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.673220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.673253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.673281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.673301 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.776085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.776184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.776222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.776258 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.776282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.878820 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.878885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.878898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.878919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.878934 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.981867 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.981915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.981927 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.981947 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.981960 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.992342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.992395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.992419 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.992445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:23 crc kubenswrapper[4726]: I1004 03:41:23.992466 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:23Z","lastTransitionTime":"2025-10-04T03:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.009298 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.013553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.013582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.013595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.013611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.013622 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.031193 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.035670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.035704 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.035716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.035731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.035745 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.049969 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.054272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.054330 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.054353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.054384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.054409 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.074224 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.078457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.078503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.078519 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.078538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.078553 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.094937 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.095159 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.096570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.096641 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.096667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.096696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.096723 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.199295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.199352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.199368 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.199391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.199411 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.303051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.303163 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.303193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.303221 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.303239 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.406786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.406971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.407000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.407035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.407060 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.502634 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:24 crc kubenswrapper[4726]: E1004 03:41:24.502845 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.510813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.510894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.510919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.510952 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.510974 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.614514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.614562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.614577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.614595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.614607 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.717070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.717141 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.717150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.717165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.717175 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.819468 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.819533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.819552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.819575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.819594 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.869228 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.898465 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.915355 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.922086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.922148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.922160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.922176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.922188 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:24Z","lastTransitionTime":"2025-10-04T03:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.928838 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.939295 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.951504 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.963870 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.977062 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:24 crc kubenswrapper[4726]: I1004 03:41:24.993271 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:24Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.009262 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.022703 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.023945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.023991 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.024068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.024089 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.024124 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.034631 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.050646 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.069086 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.085742 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.098363 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.108848 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.123927 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:25Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.126614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.126648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.126658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.126672 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.126683 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.229393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.229429 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.229437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.229452 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.229461 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.332559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.332593 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.332604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.332617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.332626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.435492 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.435538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.435552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.435570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.435582 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.501763 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.501804 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:25 crc kubenswrapper[4726]: E1004 03:41:25.502200 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.501950 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:25 crc kubenswrapper[4726]: E1004 03:41:25.502365 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:25 crc kubenswrapper[4726]: E1004 03:41:25.502531 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.538096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.538161 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.538173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.538189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.538200 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.641042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.641632 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.641726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.641806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.641864 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.744686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.745097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.745359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.745537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.745720 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.848004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.848073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.848086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.848134 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.848149 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.950722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.950787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.950837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.950861 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:25 crc kubenswrapper[4726]: I1004 03:41:25.950879 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:25Z","lastTransitionTime":"2025-10-04T03:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.054125 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.054160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.054170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.054184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.054216 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.157973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.158032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.158049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.158079 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.158096 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.260700 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.260749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.260773 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.260801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.260824 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.365025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.365250 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.365297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.365329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.365380 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.468415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.468500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.468524 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.468555 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.468579 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.502309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:26 crc kubenswrapper[4726]: E1004 03:41:26.502549 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.528493 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.551180 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.571422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.571491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.571516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.571545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.571567 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.575068 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.594733 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.611041 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.625200 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.639912 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.656232 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.672516 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.674992 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.675029 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.675037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.675051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.675063 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.686615 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.703938 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.726161 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.741327 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.756797 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.774306 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.777027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.777094 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.777126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.777144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.777156 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.790706 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.805013 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:26Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.879453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.879518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.879541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.879567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.879588 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.983198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.983269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.983295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.983323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:26 crc kubenswrapper[4726]: I1004 03:41:26.983345 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:26Z","lastTransitionTime":"2025-10-04T03:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.086310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.086379 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.086396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.086413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.086437 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.189640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.190008 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.190026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.190168 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.190189 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.293990 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.294035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.294053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.294075 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.294091 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.396824 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.396873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.396889 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.396912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.396933 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.499566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.499623 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.499640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.499661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.499676 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.502142 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.502174 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.502252 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:27 crc kubenswrapper[4726]: E1004 03:41:27.502523 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:27 crc kubenswrapper[4726]: E1004 03:41:27.502648 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:27 crc kubenswrapper[4726]: E1004 03:41:27.502738 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.602551 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.602621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.602648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.602683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.602707 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.705631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.705685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.705702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.705724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.705743 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.808298 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.808367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.808384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.808409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.808427 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.910971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.911034 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.911051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.911076 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:27 crc kubenswrapper[4726]: I1004 03:41:27.911093 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:27Z","lastTransitionTime":"2025-10-04T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.014189 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.014277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.014300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.014329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.014355 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.117787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.117849 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.117866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.117890 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.117908 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.221215 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.221287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.221303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.221324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.221333 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.325238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.325305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.325316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.325333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.325346 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.429193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.429252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.429270 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.429292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.429309 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.502293 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:28 crc kubenswrapper[4726]: E1004 03:41:28.502792 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.531972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.532068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.532093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.532202 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.532230 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.636374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.636443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.636471 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.636502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.636526 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.739425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.739487 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.739504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.739529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.739546 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.842742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.842795 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.842858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.842883 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.843488 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.945747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.945828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.945845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.945866 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:28 crc kubenswrapper[4726]: I1004 03:41:28.945884 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:28Z","lastTransitionTime":"2025-10-04T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.049235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.049314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.049329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.049352 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.049369 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.152623 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.152688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.152711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.152744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.152770 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.255732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.255802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.255815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.255837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.255853 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.358884 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.358975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.358989 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.359005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.359018 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.462165 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.462227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.462244 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.462268 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.462285 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.502073 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.502132 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:29 crc kubenswrapper[4726]: E1004 03:41:29.502249 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.502276 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:29 crc kubenswrapper[4726]: E1004 03:41:29.502389 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:29 crc kubenswrapper[4726]: E1004 03:41:29.502469 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.578058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.578188 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.578248 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.578279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.578349 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.682068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.682160 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.682178 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.682218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.682238 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.784536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.784600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.784620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.784647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.784667 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.886546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.886624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.886642 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.886666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.886684 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.990044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.990096 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.990131 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.990152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:29 crc kubenswrapper[4726]: I1004 03:41:29.990168 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:29Z","lastTransitionTime":"2025-10-04T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.092734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.092794 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.092812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.092835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.092854 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.196159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.196233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.196254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.196278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.196296 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.299181 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.299224 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.299237 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.299257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.299268 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.402278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.402345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.402363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.402392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.402411 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.502362 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:30 crc kubenswrapper[4726]: E1004 03:41:30.502506 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.504730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.504771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.504783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.504799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.504812 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.607969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.608022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.608038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.608057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.608072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.711090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.711159 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.711177 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.711204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.711221 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.814567 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.814654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.814664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.814681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.814691 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.917933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.918068 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.918100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.918175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:30 crc kubenswrapper[4726]: I1004 03:41:30.918197 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:30Z","lastTransitionTime":"2025-10-04T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.021530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.021612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.021649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.021681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.021712 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.125191 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.125262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.125285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.125314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.125336 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.233591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.233664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.233763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.233805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.233832 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.336918 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.337028 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.337046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.337071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.337088 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.439712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.439778 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.439790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.439807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.439822 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.501678 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.501727 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:31 crc kubenswrapper[4726]: E1004 03:41:31.501889 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.501937 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:31 crc kubenswrapper[4726]: E1004 03:41:31.501995 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:31 crc kubenswrapper[4726]: E1004 03:41:31.502225 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.542285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.542377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.542395 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.542417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.542435 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.644683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.644768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.644793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.644828 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.644852 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.747155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.747227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.747238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.747257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.747270 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.849535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.849599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.849612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.849630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.849646 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.953199 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.953256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.953279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.953306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:31 crc kubenswrapper[4726]: I1004 03:41:31.953327 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:31Z","lastTransitionTime":"2025-10-04T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.056752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.056823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.056856 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.056886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.056907 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.159823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.159879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.159892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.159911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.159923 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.262218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.262253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.262264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.262278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.262289 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.365062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.365098 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.365132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.365146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.365156 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.467578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.467621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.467631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.467644 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.467654 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.502352 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:32 crc kubenswrapper[4726]: E1004 03:41:32.502908 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.503174 4726 scope.go:117] "RemoveContainer" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" Oct 04 03:41:32 crc kubenswrapper[4726]: E1004 03:41:32.503452 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.569631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.569678 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.569688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.569702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.569711 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.672267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.672316 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.672333 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.672358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.672377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.774710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.774738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.774748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.774763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.774773 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.876730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.876771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.876782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.876798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.876809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.978837 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.978892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.978928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.978945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:32 crc kubenswrapper[4726]: I1004 03:41:32.978954 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:32Z","lastTransitionTime":"2025-10-04T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.080562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.080584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.080592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.080603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.080610 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.182491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.182538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.182554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.182574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.182590 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.285600 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.285658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.285676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.285705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.285723 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.388694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.388746 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.388759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.388776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.388792 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.490744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.490784 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.490796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.490812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.490828 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.502255 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.502272 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.502272 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:33 crc kubenswrapper[4726]: E1004 03:41:33.502610 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:33 crc kubenswrapper[4726]: E1004 03:41:33.502691 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:33 crc kubenswrapper[4726]: E1004 03:41:33.502413 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.593217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.593323 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.593349 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.593377 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.593400 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.696302 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.696353 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.696370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.696391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.696410 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.803291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.803523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.803647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.803744 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.803832 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.906490 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.906531 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.906549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.906572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:33 crc kubenswrapper[4726]: I1004 03:41:33.906591 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:33Z","lastTransitionTime":"2025-10-04T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.008925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.008986 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.008996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.009044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.009055 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.111152 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.111211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.111230 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.111256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.111274 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.213307 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.213339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.213347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.213364 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.213373 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.316638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.316676 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.316686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.316702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.316711 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.349848 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.349943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.349969 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.349999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.350021 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.365327 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.372977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.373009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.373021 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.373039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.373052 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.385501 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.390391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.390422 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.390433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.390465 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.390477 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.402758 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.407259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.407520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.407648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.407791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.407934 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.421793 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.425967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.426006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.426016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.426032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.426042 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.438900 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:34Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.439503 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.441370 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.441435 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.441453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.441479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.441496 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.502342 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.502557 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.543628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.543660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.543668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.543681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.543691 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.645597 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.645688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.645715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.645748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.645770 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.748657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.748899 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.748959 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.749026 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.749086 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.754217 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.754356 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:34 crc kubenswrapper[4726]: E1004 03:41:34.754414 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:06.754397834 +0000 UTC m=+100.929021047 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.851423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.851467 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.851481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.851498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.851510 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.954772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.954850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.954864 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.954902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:34 crc kubenswrapper[4726]: I1004 03:41:34.954917 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:34Z","lastTransitionTime":"2025-10-04T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.057282 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.057325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.057337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.057354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.057366 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.159748 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.159825 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.159850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.159891 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.159908 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.262893 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.262962 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.262980 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.263004 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.263021 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.365549 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.365610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.365629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.365652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.365669 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.468005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.468146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.468171 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.468203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.468225 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.501400 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:35 crc kubenswrapper[4726]: E1004 03:41:35.501562 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.501443 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:35 crc kubenswrapper[4726]: E1004 03:41:35.501742 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.501400 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:35 crc kubenswrapper[4726]: E1004 03:41:35.501944 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.571396 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.571432 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.571443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.571459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.571470 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.674294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.674636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.674785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.674967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.675097 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.778151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.778217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.778238 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.778264 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.778283 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.880894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.880941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.880957 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.880993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.881009 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.983882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.983946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.983971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.983998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:35 crc kubenswrapper[4726]: I1004 03:41:35.984031 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:35Z","lastTransitionTime":"2025-10-04T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.085708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.086041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.086186 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.086334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.086448 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.189001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.189030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.189038 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.189051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.189060 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.291822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.291896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.291913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.291937 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.291954 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.394016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.394041 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.394049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.394061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.394070 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.496912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.496950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.496964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.496984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.496998 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.501962 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:36 crc kubenswrapper[4726]: E1004 03:41:36.502141 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.520726 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.541635 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.558876 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.571980 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.587218 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.598791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.598950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.599030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.599142 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.599282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.605086 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.646203 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.665719 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.677723 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.690773 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.701350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.701401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.701418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.701437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.701449 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.705904 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.718335 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.730975 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.747779 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.761331 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.776946 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.794264 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:36Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.804195 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.804247 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.804262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.804287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.804304 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.906336 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.906367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.906376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.906391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.906401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:36Z","lastTransitionTime":"2025-10-04T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.996709 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/0.log" Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.996763 4726 generic.go:334] "Generic (PLEG): container finished" podID="7fa9344a-8c9c-4bc5-b91e-b6c1400d4025" containerID="6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd" exitCode=1 Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.996815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerDied","Data":"6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd"} Oct 04 03:41:36 crc kubenswrapper[4726]: I1004 03:41:36.997527 4726 scope.go:117] "RemoveContainer" containerID="6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.008919 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.008946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.008956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.008968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.008977 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.010879 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.025729 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.044971 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.062125 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.072091 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.083157 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.092863 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.105815 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.111441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.111474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.111484 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.111502 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.111515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.115020 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.123969 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.134066 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.148352 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.162681 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.181031 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.192892 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.206133 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.214090 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.214144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.214154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.214169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.214180 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.218548 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:37Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.316425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.316469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.316480 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.316498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.316510 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.418577 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.418634 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.418647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.418663 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.418676 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.502192 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.502225 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.502312 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:37 crc kubenswrapper[4726]: E1004 03:41:37.502439 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:37 crc kubenswrapper[4726]: E1004 03:41:37.502615 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:37 crc kubenswrapper[4726]: E1004 03:41:37.502699 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.520560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.520591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.520603 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.520619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.520631 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.622863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.622897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.622908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.622924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.622935 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.724757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.724786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.724797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.724813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.724824 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.826694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.826747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.826760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.826779 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.826793 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.929482 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.929541 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.929558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.929582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:37 crc kubenswrapper[4726]: I1004 03:41:37.929600 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:37Z","lastTransitionTime":"2025-10-04T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.002213 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/0.log" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.002314 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerStarted","Data":"f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.015615 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.029126 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.031953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.032024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.032037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.032077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.032090 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.039684 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.047938 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.057719 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.069299 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.082777 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.101019 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.115029 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.126434 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.134570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.134610 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.134622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.134639 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.134650 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.138089 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.148872 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.158827 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.171023 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.190758 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.204488 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.214690 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:38Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.237614 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.237648 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.237659 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.237675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.237689 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.340246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.340283 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.340291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.340319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.340329 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.443449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.443538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.443561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.443590 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.443612 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.502612 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:38 crc kubenswrapper[4726]: E1004 03:41:38.503058 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.521052 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.546570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.546745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.546829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.546913 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.547034 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.649010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.649046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.649057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.649072 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.649082 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.751144 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.751182 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.751197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.751217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.751231 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.852729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.852776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.852788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.852804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.852816 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.955402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.955438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.955448 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.955462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:38 crc kubenswrapper[4726]: I1004 03:41:38.955470 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:38Z","lastTransitionTime":"2025-10-04T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.057613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.057657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.057670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.057687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.057699 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.160592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.160631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.160643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.160658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.160669 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.262173 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.262205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.262214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.262226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.262236 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.364653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.364685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.364693 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.364707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.364715 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.467747 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.467783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.467791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.467809 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.467818 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.502132 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.502311 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.502334 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:39 crc kubenswrapper[4726]: E1004 03:41:39.502410 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:39 crc kubenswrapper[4726]: E1004 03:41:39.502312 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:39 crc kubenswrapper[4726]: E1004 03:41:39.502526 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.569717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.569766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.569783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.569806 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.569824 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.673000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.673097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.673162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.673196 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.673232 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.775897 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.775943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.775955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.775973 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.775986 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.877912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.877950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.877961 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.877977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.877989 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.980642 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.980805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.980823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.980850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:39 crc kubenswrapper[4726]: I1004 03:41:39.980862 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:39Z","lastTransitionTime":"2025-10-04T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.083734 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.083775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.083786 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.083800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.083809 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.186401 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.186458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.186476 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.186503 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.186523 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.288661 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.288698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.288709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.288724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.288735 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.391640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.391684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.391695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.391714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.391726 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.494459 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.494499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.494509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.494523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.494534 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.501959 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:40 crc kubenswrapper[4726]: E1004 03:41:40.502076 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.597523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.597561 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.597572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.597591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.597603 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.699968 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.700009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.700020 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.700036 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.700046 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.803012 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.803051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.803061 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.803076 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.803087 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.905847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.905892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.905902 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.905916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:40 crc kubenswrapper[4726]: I1004 03:41:40.905925 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:40Z","lastTransitionTime":"2025-10-04T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.008912 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.009175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.009184 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.009197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.009206 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.112179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.112218 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.112225 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.112241 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.112249 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.214517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.214558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.214566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.214579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.214588 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.317023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.317100 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.317154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.317183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.317203 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.419946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.419975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.419982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.419994 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.420003 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.501907 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:41 crc kubenswrapper[4726]: E1004 03:41:41.502026 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.501917 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.502061 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:41 crc kubenswrapper[4726]: E1004 03:41:41.502090 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:41 crc kubenswrapper[4726]: E1004 03:41:41.502229 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.521851 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.521900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.521916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.521940 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.522004 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.624392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.624437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.624446 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.624460 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.624469 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.727458 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.727529 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.727547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.727570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.727585 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.833570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.833613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.833628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.833651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.833667 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.936203 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.936254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.936266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.936285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:41 crc kubenswrapper[4726]: I1004 03:41:41.936298 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:41Z","lastTransitionTime":"2025-10-04T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.038763 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.038807 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.038819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.038835 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.038845 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.141798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.141862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.141879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.141903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.141920 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.246027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.246073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.246086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.246102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.246130 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.349053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.349183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.349207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.349235 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.349252 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.452194 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.452266 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.452291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.452319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.452341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.502341 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:42 crc kubenswrapper[4726]: E1004 03:41:42.502524 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.555267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.555343 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.555359 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.555384 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.555401 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.658206 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.658254 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.658277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.658305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.658328 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.760983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.761009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.761019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.761033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.761041 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.864031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.864439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.864651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.864845 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.865032 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.968711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.968776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.968799 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.968829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:42 crc kubenswrapper[4726]: I1004 03:41:42.968853 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:42Z","lastTransitionTime":"2025-10-04T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.071240 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.071303 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.071322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.071344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.071361 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.174813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.174909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.174928 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.174955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.174977 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.278179 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.278252 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.278275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.278306 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.278330 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.381606 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.381660 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.381672 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.381690 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.381704 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.484318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.484391 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.484409 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.484434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.484452 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.502048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.502051 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.502337 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:43 crc kubenswrapper[4726]: E1004 03:41:43.502468 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:43 crc kubenswrapper[4726]: E1004 03:41:43.502827 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:43 crc kubenswrapper[4726]: E1004 03:41:43.502958 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.586967 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.587037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.587062 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.587093 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.587147 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.690649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.690701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.690724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.690755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.690775 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.793239 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.793389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.793420 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.793447 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.793468 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.896421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.896479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.896498 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.896525 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.896547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.999880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.999949 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:43 crc kubenswrapper[4726]: I1004 03:41:43.999972 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:43.999999 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.000019 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:43Z","lastTransitionTime":"2025-10-04T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.104045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.104132 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.104156 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.104185 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.104235 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.208025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.208085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.208102 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.208176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.208199 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.310882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.310956 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.310979 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.311010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.311030 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.414619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.414697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.414727 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.414760 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.414786 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.502467 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.503214 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.503512 4726 scope.go:117] "RemoveContainer" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.517921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.517981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.518000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.518027 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.518044 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.620924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.620978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.620996 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.621021 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.621039 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.663757 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.663792 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.663800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.663813 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.663821 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.679527 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.685594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.685635 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.685657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.685684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.685708 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.706402 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.711575 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.711656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.711682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.711716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.711741 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.731549 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.736205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.736275 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.736294 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.736322 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.736342 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.756069 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.761553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.761620 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.761640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.761668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.761689 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.780087 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:44Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:44 crc kubenswrapper[4726]: E1004 03:41:44.780274 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.782440 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.782500 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.782514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.782533 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.782546 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.885133 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.885687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.886192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.886297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.887377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.990019 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.990045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.990070 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.990082 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:44 crc kubenswrapper[4726]: I1004 03:41:44.990090 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:44Z","lastTransitionTime":"2025-10-04T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.044087 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/2.log" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.048947 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.049580 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.050093 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.064702 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.078526 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6aa2546-1150-4858-a964-904cb977bca2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b79980747936da340dcc0b0fecc9d02f5802535b81766bcdb254ecbfb44fc835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092126 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092229 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.092297 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.111294 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.126612 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.137305 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.147638 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.163983 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.178267 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.188920 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.198640 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.198689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.198707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.198722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.198731 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.205368 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.216077 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.225995 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.238801 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.249981 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.262709 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.273380 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.286407 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:45Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.301094 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.301140 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.301149 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.301170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.301178 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.402887 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.402923 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.402933 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.402945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.402954 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.501975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.502007 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.502148 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:45 crc kubenswrapper[4726]: E1004 03:41:45.502194 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:45 crc kubenswrapper[4726]: E1004 03:41:45.502211 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:45 crc kubenswrapper[4726]: E1004 03:41:45.502259 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.505537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.505578 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.505594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.505645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.505664 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.609453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.609520 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.609538 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.609564 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.609586 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.712910 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.712951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.712965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.712981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.712992 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.815725 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.815796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.815819 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.815850 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.815872 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.919375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.919437 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.919462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.919494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:45 crc kubenswrapper[4726]: I1004 03:41:45.919515 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:45Z","lastTransitionTime":"2025-10-04T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.022637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.022688 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.022706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.022728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.022744 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.055604 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.056419 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/2.log" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.061240 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.062279 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" exitCode=1 Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.062329 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.062371 4726 scope.go:117] "RemoveContainer" containerID="14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.063578 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:41:46 crc kubenswrapper[4726]: E1004 03:41:46.063942 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.081665 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.101404 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.119400 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.126731 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.126781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.126798 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.126822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.126841 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.151204 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:45Z\\\",\\\"message\\\":\\\"er-operator template LB for network=default: []services.LB{}\\\\nI1004 03:41:45.571419 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:41:45.571470 6751 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nI1004 03:41:45.571492 6751 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nF1004 03:41:45.571496 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.178344 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.199269 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.218265 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.228984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.229054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.229071 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.229095 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.229149 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.240957 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.262046 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.281841 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.302047 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.324759 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.332193 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.332297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.332319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.332345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.332364 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.346049 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.371011 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.390967 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.406612 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.425236 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.437331 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.437417 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.437443 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.437475 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.437499 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.442004 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6aa2546-1150-4858-a964-904cb977bca2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b79980747936da340dcc0b0fecc9d02f5802535b81766bcdb254ecbfb44fc835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.502023 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:46 crc kubenswrapper[4726]: E1004 03:41:46.502200 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.517392 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.535045 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.540618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.540662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.540679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.540701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.540719 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.553410 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.568148 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.582092 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.592729 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.603547 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.613758 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.623813 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6aa2546-1150-4858-a964-904cb977bca2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b79980747936da340dcc0b0fecc9d02f5802535b81766bcdb254ecbfb44fc835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.642126 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.642162 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.642180 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.642198 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.642206 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.643187 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.656320 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.668895 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.683298 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.695637 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.710304 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.733521 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14ba19a50c447631686b0446fbb4bf4da4da3efd55cc3e096782c166e7701093\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:20Z\\\",\\\"message\\\":\\\"s/informers/externalversions/factory.go:140\\\\nI1004 03:41:20.382237 6416 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 03:41:20.382298 6416 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.382354 6416 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 03:41:20.388614 6416 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 03:41:20.388776 6416 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 03:41:20.388939 6416 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 03:41:20.389059 6416 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 03:41:20.389092 6416 factory.go:656] Stopping watch factory\\\\nI1004 03:41:20.418489 6416 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 03:41:20.418521 6416 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 03:41:20.418581 6416 ovnkube.go:599] Stopped ovnkube\\\\nI1004 03:41:20.418619 6416 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 03:41:20.418711 6416 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:45Z\\\",\\\"message\\\":\\\"er-operator template LB for network=default: []services.LB{}\\\\nI1004 03:41:45.571419 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:41:45.571470 6751 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nI1004 03:41:45.571492 6751 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nF1004 03:41:45.571496 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.744518 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.744543 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.744552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.744570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.744582 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.750923 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.767190 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:46Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.847971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.848025 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.848035 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.848055 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.848065 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.950953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.951016 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.951031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.951054 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:46 crc kubenswrapper[4726]: I1004 03:41:46.951072 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:46Z","lastTransitionTime":"2025-10-04T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.055943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.056363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.056558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.056718 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.056859 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.068759 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.074448 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.080056 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:41:47 crc kubenswrapper[4726]: E1004 03:41:47.080404 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.097610 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-95wxp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb0a03c5-46d9-4815-ab49-a93336faeff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c6fc53dbbfff983912d96e96b8722e251ae70910d6ed9e521b62a9778bd9af0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l5q2w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-95wxp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.114481 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13f6cac4-0dd4-466d-820f-0822d4e0128f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c82e7c6d57f562913bf38a449f10e7104c34c9764c332fb6fff6a7c4841b4c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02f2a440c02c661b2f814b4c9b80083f5dfd0dc594cb24c426b0f5447de760d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xr6f9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xvgfd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.150387 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8d6c450-3310-4058-90dc-0a39ab1934c4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"message\\\":\\\"++ K8S_NODE=\\\\n++ [[ -n '' ]]\\\\n++ northd_pidfile=/var/run/ovn/ovn-northd.pid\\\\n++ controller_pidfile=/var/run/ovn/ovn-controller.pid\\\\n++ controller_logfile=/var/log/ovn/acl-audit-log.log\\\\n++ vswitch_dbsock=/var/run/openvswitch/db.sock\\\\n++ nbdb_pidfile=/var/run/ovn/ovnnb_db.pid\\\\n++ nbdb_sock=/var/run/ovn/ovnnb_db.sock\\\\n++ nbdb_ctl=/var/run/ovn/ovnnb_db.ctl\\\\n++ sbdb_pidfile=/var/run/ovn/ovnsb_db.pid\\\\n++ sbdb_sock=/var/run/ovn/ovnsb_db.sock\\\\n++ sbdb_ctl=/var/run/ovn/ovnsb_db.ctl\\\\n+ start-audit-log-rotation\\\\n+ MAXFILESIZE=50000000\\\\n+ MAXLOGFILES=5\\\\n++ dirname /var/log/ovn/acl-audit-log.log\\\\n+ LOGDIR=/var/log/ovn\\\\n+ local retries=0\\\\n+ [[ 30 -gt 0 ]]\\\\n+ (( retries += 1 ))\\\\n++ cat /var/run/ovn/ovn-controller.pid\\\\ncat: /var/run/ovn/ovn-controller.pid: No such file or directory\\\\n+ CONTROLLERPID=\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:45Z\\\",\\\"message\\\":\\\"er-operator template LB for network=default: []services.LB{}\\\\nI1004 03:41:45.571419 6751 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 03:41:45.571470 6751 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nI1004 03:41:45.571492 6751 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-9777l\\\\nF1004 03:41:45.571496 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:41:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xqlqr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-snjd6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.160683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.160758 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.160783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.160812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.160832 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.175506 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t24l4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fab2e099-b326-498e-bd92-96260472132c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5aa2f8001913b01aaca405fd7eafd68c8b59b760a28675fdd75588899a38aca0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47520b952ed4cbd648ea0196eb8c71f47ec7755d68af6fcace8226388a9c4b35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef12a05a09a38fc613b44ccbbe3df1106b6829c2f2be5f220c408522c975a1f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b776e68d358ad00f4c905767106d9e58a3c325861b2430809b5648d1a12a79db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f557e02e76a5d0f44808b562343855242a17a66ee60c951cdea1d3f09af9c44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc3de76f73ee6fa493477b3146fc9b14fee638803f7c9182a54e4ab8220ecab6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ef0b1a06e02d796a37ea0311f585a7865f3afba19dbe79ac1bae38abf1c2cc6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xfzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t24l4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.194079 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-b9rkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32c5bb99-f0d4-4259-8604-5b8664e20a5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9fd0f2242d2fe195a26f0690401562defdc431b19b5ff7dc847f32e5bfda00a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbkw9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-b9rkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.212450 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://383e9d92eb6f8f3ce6c5c25a1dbb3c4c12b7d9d13d8ed08e55727afdf71935be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4kv8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9777l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.229277 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9qn78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77b969b7-0698-4589-a6cf-c08cc779ffe2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22dkh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:41:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9qn78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.254070 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25bac92a-f4f9-4fc8-bc62-b551233fc334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1da3f58a8d2aa1f2aa5895e7913ffe1fd51c4df45b7b824b50d386494fa7eeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://51eb8b9fdde849ed9b92e601ff88e0963e93bb0207252568108f83167681ea65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fb878b383fe277a66329e5e38e75da5dc8f16f1895c946726f1f063f906d8b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://568279f7938c79bbe07d864de51cf37a0c05ee565d9f602b3ffc49611e67e8a2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.263697 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.263844 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.263945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.264064 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.264245 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.269271 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511ad39d-2966-46a8-a99b-225d87f6705a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2e98e2a87f9d9e097b01f4a6abc8f927d66e3ec1d8049d7daa5fe0d3b72c594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3ebabe3f1c2731bded8c2d95b6f9c02f1a16f0907a2ee806a69178622531c10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e66c95bad48a146741487b9cfe59506399593cb051afce1f12d68f836fbdeca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a8ce88e2007918846a3fd81fc2d7a3fdfe80767c0e3d0ca5d4b66cab3dc8162\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.285986 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b07414b6e23b86a0237a48f798309ebf8b256460a8eab79fd1534d2474d36379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.301042 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.319739 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.336444 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04118b219f9328279710b2705e5e729a09aac422a2b62ca9000edbea8f48fae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42edb45accfe7405400bfbed65dbb7de150c061afccc4c50356d41bb4ec7d18c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.349198 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6aa2546-1150-4858-a964-904cb977bca2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b79980747936da340dcc0b0fecc9d02f5802535b81766bcdb254ecbfb44fc835\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3495e04948dfe1934028219dbaf021d9dde4ea45f0f300f62b4f45c2bb21998\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.366673 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.366709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.366721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.366737 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.366749 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.368281 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5bcfc2-c40d-4fdc-b253-5733197eeab4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://154c856b8f760f569d49e7ebaac10df8394727ee391ebe0d79375e6f881b5eb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0837a2c1c69547042803f53616ab4c0c7fab2f3e7dc238bbbaf6465f922f3b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30dd1d850cda011f018caf5f794b20a64a66d1d30b8600e48085a3973a316e4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d49a49317e10e9082c76f8297c261af763164823e2956e0712cc30008f6ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fab0b9393c787be3a06341ba74e1d9487f083d531f5c54c47da8403025675fa\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 03:40:46.706311 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 03:40:46.706539 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 03:40:46.707677 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3267858508/tls.crt::/tmp/serving-cert-3267858508/tls.key\\\\\\\"\\\\nI1004 03:40:47.041135 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 03:40:47.050265 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 03:40:47.050317 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 03:40:47.050359 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 03:40:47.050373 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 03:40:47.061912 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1004 03:40:47.061963 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061975 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 03:40:47.061986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 03:40:47.061994 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 03:40:47.062001 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 03:40:47.062007 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 03:40:47.062014 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1004 03:40:47.064933 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:40Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://274b9d2381f01f7f3da94714ef906fccdf1169565bbde03616178bb5d25d4af0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e3b8a4fd149132f37d96e85fc47bcdde219bb6b47cdb05d265e5bb911d5d15c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T03:40:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.385420 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:47Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.400200 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba14fa20f39ce0bf9b35d59ae0d82bce38d2e00d0b7a1bee09f7cef38f00d962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:40:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.419580 4726 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5c96v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:40:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T03:41:36Z\\\",\\\"message\\\":\\\"2025-10-04T03:40:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f\\\\n2025-10-04T03:40:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_21363561-7177-49f4-b245-95f7577e341f to /host/opt/cni/bin/\\\\n2025-10-04T03:40:51Z [verbose] multus-daemon started\\\\n2025-10-04T03:40:51Z [verbose] Readiness Indicator file check\\\\n2025-10-04T03:41:36Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T03:40:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T03:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8bxc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T03:40:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5c96v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:47Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.469479 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.469539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.469558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.469583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.469602 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.501728 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.501812 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:47 crc kubenswrapper[4726]: E1004 03:41:47.501945 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:47 crc kubenswrapper[4726]: E1004 03:41:47.502200 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.502232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:47 crc kubenswrapper[4726]: E1004 03:41:47.502398 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.572376 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.572415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.572423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.572441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.572452 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.675742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.675789 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.675802 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.675822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.675836 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.779408 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.779495 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.779521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.779554 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.779580 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.885161 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.885227 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.885245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.885272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.885292 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.988315 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.988374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.988397 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.988423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:47 crc kubenswrapper[4726]: I1004 03:41:47.988440 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:47Z","lastTransitionTime":"2025-10-04T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.091211 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.091274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.091292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.091319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.091336 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.194277 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.194339 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.194357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.194382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.194399 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.298385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.298466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.298491 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.298523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.298547 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.401812 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.402022 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.402043 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.402300 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.402326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.502527 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:48 crc kubenswrapper[4726]: E1004 03:41:48.502748 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.505527 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.505591 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.505619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.505651 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.505675 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.609074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.609164 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.609183 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.609209 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.609228 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.712281 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.712341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.712360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.712392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.712411 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.815169 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.815568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.815592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.815622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.815643 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.918685 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.918954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.919053 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.919192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:48 crc kubenswrapper[4726]: I1004 03:41:48.919290 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:48Z","lastTransitionTime":"2025-10-04T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.021658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.021728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.021752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.021783 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.021808 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.124537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.124588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.124605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.124629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.124648 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.227350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.227427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.227445 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.227470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.227491 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.331037 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.331092 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.331139 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.331170 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.331194 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.434329 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.434383 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.434402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.434425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.434441 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.502473 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.502470 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:49 crc kubenswrapper[4726]: E1004 03:41:49.502650 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:49 crc kubenswrapper[4726]: E1004 03:41:49.502805 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.502502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:49 crc kubenswrapper[4726]: E1004 03:41:49.502917 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.537024 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.537080 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.537097 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.537146 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.537172 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.639724 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.639766 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.639776 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.639791 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.639801 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.743461 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.743823 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.744015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.744214 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.744393 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.847466 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.847536 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.847560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.847587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.847609 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.950894 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.951390 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.951615 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.951832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:49 crc kubenswrapper[4726]: I1004 03:41:49.952050 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:49Z","lastTransitionTime":"2025-10-04T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.055539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.055900 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.056086 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.056319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.056512 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.159382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.159427 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.159439 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.159456 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.159467 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.262611 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.262670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.262699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.262723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.262740 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.365585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.365677 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.365699 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.365726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.365749 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.468702 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.468750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.468759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.468772 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.468780 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.502546 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:50 crc kubenswrapper[4726]: E1004 03:41:50.502674 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.572226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.572279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.572296 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.572319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.572336 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.675505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.675562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.675579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.675601 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.675618 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.778504 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.778570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.778596 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.778628 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.778650 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.881909 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.881993 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.882018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.882051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.882075 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.984901 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.984955 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.984971 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.984997 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:50 crc kubenswrapper[4726]: I1004 03:41:50.985015 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:50Z","lastTransitionTime":"2025-10-04T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.087671 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.087728 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.087745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.087769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.087786 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.191246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.191325 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.191356 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.191388 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.191413 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.299622 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.299696 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.299715 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.299822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.299911 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.402686 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.402781 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.402810 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.402841 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.402864 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.429694 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.429839 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.429879 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.429947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.429986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430192 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430220 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430239 4726 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430306 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.4302841 +0000 UTC m=+149.604907353 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430342 4726 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430429 4726 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430441 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430623 4726 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430664 4726 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430480 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.430458904 +0000 UTC m=+149.605082127 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430711 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.430698791 +0000 UTC m=+149.605322014 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430727 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.430719781 +0000 UTC m=+149.605343004 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.430739 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.430733282 +0000 UTC m=+149.605356515 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.501881 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.501972 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.502036 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.502201 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.502281 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:51 crc kubenswrapper[4726]: E1004 03:41:51.502361 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.505966 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.505995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.506006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.506023 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.506034 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.609278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.609357 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.609382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.609413 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.609437 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.713636 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.713995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.714217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.714522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.714697 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.817797 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.817858 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.817876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.817898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.817915 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.922259 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.922321 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.922337 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.922360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:51 crc kubenswrapper[4726]: I1004 03:41:51.922377 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:51Z","lastTransitionTime":"2025-10-04T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.025285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.025341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.025358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.025382 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.025403 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.128869 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.128932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.128950 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.128977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.128997 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.231607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.231674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.231687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.231711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.231726 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.335579 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.335665 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.335682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.335710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.335728 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.439267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.439341 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.439358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.439375 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.439386 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.502243 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:52 crc kubenswrapper[4726]: E1004 03:41:52.502438 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.541898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.541985 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.542005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.542032 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.542052 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.649438 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.649494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.649510 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.649535 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.649551 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.753405 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.753681 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.753749 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.753817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.753906 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.856515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.856566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.856582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.856604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.856623 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.959425 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.959775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.959925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.960262 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:52 crc kubenswrapper[4726]: I1004 03:41:52.960412 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:52Z","lastTransitionTime":"2025-10-04T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.063706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.063787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.063815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.063847 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.063871 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.167030 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.167094 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.167176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.167201 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.167219 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.270434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.270515 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.270539 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.270570 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.270592 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.373876 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.373936 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.373953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.373975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.373991 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.477385 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.477430 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.477444 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.477462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.477475 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.502241 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.502309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.502253 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:53 crc kubenswrapper[4726]: E1004 03:41:53.502440 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:53 crc kubenswrapper[4726]: E1004 03:41:53.502551 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:53 crc kubenswrapper[4726]: E1004 03:41:53.502649 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.581978 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.582033 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.582051 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.582074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.582091 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.685371 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.685433 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.685454 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.685481 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.685499 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.788537 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.788619 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.788643 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.788669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.788687 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.891815 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.891886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.891911 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.891943 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.891966 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.994870 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.994915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.994932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.994953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:53 crc kubenswrapper[4726]: I1004 03:41:53.994970 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:53Z","lastTransitionTime":"2025-10-04T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.099155 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.099233 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.099253 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.099279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.099298 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.202631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.202701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.202726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.202759 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.202781 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.306630 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.306692 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.306712 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.306755 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.306785 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.410197 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.410260 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.410279 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.410305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.410324 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.502670 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:54 crc kubenswrapper[4726]: E1004 03:41:54.502881 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.512716 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.512751 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.512768 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.512790 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.512807 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.616423 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.616474 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.616488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.616517 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.616532 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.719670 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.719762 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.719785 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.719808 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.719830 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.823921 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.823989 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.824012 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.824039 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.824057 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.907217 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.907295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.907318 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.907348 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.907365 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: E1004 03:41:54.928679 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.934222 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.934295 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.934314 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.934342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.934383 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: E1004 03:41:54.957960 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.963488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.963548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.963566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.963592 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.963611 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:54 crc kubenswrapper[4726]: E1004 03:41:54.984805 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:54Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.990548 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.990618 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.990637 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.990662 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:54 crc kubenswrapper[4726]: I1004 03:41:54.990679 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:54Z","lastTransitionTime":"2025-10-04T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.010986 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.015631 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.015722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.015750 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.015775 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.015792 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.035565 4726 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T03:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"10cbd887-f465-43af-a376-e5f244cebbdc\\\",\\\"systemUUID\\\":\\\"64e1ddb6-4f16-4c65-97b7-a8481b56f368\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T03:41:55Z is after 2025-08-24T17:21:41Z" Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.035834 4726 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.037898 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.037965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.037983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.038011 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.038030 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.141602 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.141674 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.141698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.141730 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.141753 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.244638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.244689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.244705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.244726 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.244742 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.347497 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.347553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.347572 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.347595 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.347611 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.450624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.450695 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.450708 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.450723 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.450734 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.501649 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.502041 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.502317 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.502442 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.502455 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:55 crc kubenswrapper[4726]: E1004 03:41:55.502655 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.517446 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.553534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.553669 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.553683 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.553738 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.553751 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.656885 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.656922 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.656932 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.656946 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.656958 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.759326 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.759769 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.759954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.760150 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.760326 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.863001 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.863058 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.863075 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.863099 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.863157 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.966607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.966924 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.967087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.967313 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:55 crc kubenswrapper[4726]: I1004 03:41:55.967507 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:55Z","lastTransitionTime":"2025-10-04T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.071157 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.071559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.071584 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.071609 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.071626 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.175667 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.175742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.175951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.175982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.176041 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.278941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.278984 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.278998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.279017 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.279030 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.381945 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.381998 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.382018 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.382042 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.382059 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.483854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.484362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.484470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.484546 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.484619 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.501549 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:56 crc kubenswrapper[4726]: E1004 03:41:56.501810 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.568381 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.568351602 podStartE2EDuration="18.568351602s" podCreationTimestamp="2025-10-04 03:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.530221049 +0000 UTC m=+90.704844262" watchObservedRunningTime="2025-10-04 03:41:56.568351602 +0000 UTC m=+90.742974855" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.568587 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.5685783180000001 podStartE2EDuration="1.568578318s" podCreationTimestamp="2025-10-04 03:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.565544659 +0000 UTC m=+90.740167962" watchObservedRunningTime="2025-10-04 03:41:56.568578318 +0000 UTC m=+90.743201571" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.588287 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.588345 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.588514 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.588544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.588568 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.612635 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.612607807 podStartE2EDuration="1m9.612607807s" podCreationTimestamp="2025-10-04 03:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.596076252 +0000 UTC m=+90.770699515" watchObservedRunningTime="2025-10-04 03:41:56.612607807 +0000 UTC m=+90.787231030" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.676529 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5c96v" podStartSLOduration=68.676448707 podStartE2EDuration="1m8.676448707s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.654080998 +0000 UTC m=+90.828704231" watchObservedRunningTime="2025-10-04 03:41:56.676448707 +0000 UTC m=+90.851071950" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.677744 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-95wxp" podStartSLOduration=68.677726291 podStartE2EDuration="1m8.677726291s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.675492542 +0000 UTC m=+90.850115815" watchObservedRunningTime="2025-10-04 03:41:56.677726291 +0000 UTC m=+90.852349544" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.690684 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.690988 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.691060 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.691154 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.691222 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.725436 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xvgfd" podStartSLOduration=68.725401704 podStartE2EDuration="1m8.725401704s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.696979887 +0000 UTC m=+90.871603140" watchObservedRunningTime="2025-10-04 03:41:56.725401704 +0000 UTC m=+90.900024957" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.758489 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-t24l4" podStartSLOduration=68.758465004 podStartE2EDuration="1m8.758465004s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.746361556 +0000 UTC m=+90.920984779" watchObservedRunningTime="2025-10-04 03:41:56.758465004 +0000 UTC m=+90.933088237" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.758740 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-b9rkg" podStartSLOduration=68.758734341 podStartE2EDuration="1m8.758734341s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.7579319 +0000 UTC m=+90.932555133" watchObservedRunningTime="2025-10-04 03:41:56.758734341 +0000 UTC m=+90.933357564" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.793585 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.793817 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.793970 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.794047 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.794130 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.797580 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=63.797562393 podStartE2EDuration="1m3.797562393s" podCreationTimestamp="2025-10-04 03:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.796959177 +0000 UTC m=+90.971582400" watchObservedRunningTime="2025-10-04 03:41:56.797562393 +0000 UTC m=+90.972185646" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.814937 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=39.81491711 podStartE2EDuration="39.81491711s" podCreationTimestamp="2025-10-04 03:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.814127009 +0000 UTC m=+90.988750242" watchObservedRunningTime="2025-10-04 03:41:56.81491711 +0000 UTC m=+90.989540363" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.892506 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podStartSLOduration=68.892487051 podStartE2EDuration="1m8.892487051s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:41:56.892156503 +0000 UTC m=+91.066779716" watchObservedRunningTime="2025-10-04 03:41:56.892487051 +0000 UTC m=+91.067110264" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.896745 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.896796 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.896805 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.896822 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.896831 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.999388 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.999664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.999880 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:56 crc kubenswrapper[4726]: I1004 03:41:56.999953 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.000012 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:56Z","lastTransitionTime":"2025-10-04T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.102473 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.102507 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.102516 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.102530 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.102539 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.205594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.205656 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.205675 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.205703 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.205721 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.309013 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.309087 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.309389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.309710 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.309776 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.413187 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.413257 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.413278 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.413308 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.413330 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.502524 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.502612 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:57 crc kubenswrapper[4726]: E1004 03:41:57.502737 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:57 crc kubenswrapper[4726]: E1004 03:41:57.502952 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.503295 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:57 crc kubenswrapper[4726]: E1004 03:41:57.503441 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.515804 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.515854 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.515870 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.515896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.515914 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.619236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.619290 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.619309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.619334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.619350 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.723175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.723369 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.723392 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.723415 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.723433 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.826489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.826544 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.826562 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.826583 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.826600 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.930263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.930342 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.930365 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.930393 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:57 crc kubenswrapper[4726]: I1004 03:41:57.930414 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:57Z","lastTransitionTime":"2025-10-04T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.033534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.033594 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.033613 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.033638 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.033655 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.136261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.136334 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.136360 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.136412 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.136435 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.239015 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.239063 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.239085 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.239137 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.239161 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.342558 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.342608 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.342624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.342647 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.342665 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.445488 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.445553 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.445576 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.445605 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.445625 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.502357 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:41:58 crc kubenswrapper[4726]: E1004 03:41:58.502500 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.548506 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.548565 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.548574 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.548587 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.548596 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.651709 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.651752 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.651764 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.651782 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.651794 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.754607 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.755005 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.755219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.755388 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.755540 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.859387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.859453 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.859472 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.859499 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.859520 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.963299 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.963358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.963389 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.963418 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:58 crc kubenswrapper[4726]: I1004 03:41:58.963439 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:58Z","lastTransitionTime":"2025-10-04T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.066599 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.066975 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.067060 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.067176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.067269 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.170787 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.170868 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.170892 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.170925 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.170948 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.274771 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.275220 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.275402 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.275534 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.275671 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.379896 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.379964 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.379982 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.380010 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.380034 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.483722 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.484046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.484291 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.484442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.484598 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.501550 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:41:59 crc kubenswrapper[4726]: E1004 03:41:59.501967 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.501604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:41:59 crc kubenswrapper[4726]: E1004 03:41:59.502437 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.501565 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:41:59 crc kubenswrapper[4726]: E1004 03:41:59.502808 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.587434 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.587788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.587884 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.587983 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.588085 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.692455 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.692505 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.692521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.692542 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.692559 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.796289 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.796344 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.796361 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.796387 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.796584 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.900410 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.900657 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.900679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.900706 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:41:59 crc kubenswrapper[4726]: I1004 03:41:59.900724 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:41:59Z","lastTransitionTime":"2025-10-04T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.003965 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.004031 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.004049 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.004073 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.004093 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.111077 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.111204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.111232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.111265 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.111294 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.214046 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.214367 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.214450 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.214560 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.214680 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.317398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.317449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.317457 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.317469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.317479 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.419006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.419263 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.419332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.419414 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.419493 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.502514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:00 crc kubenswrapper[4726]: E1004 03:42:00.502684 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.522354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.522421 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.522442 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.522469 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.522492 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.625207 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.625269 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.625292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.625320 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.625341 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.727882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.727941 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.727958 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.727981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.727998 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.830568 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.830646 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.830664 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.830691 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.830709 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.933732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.933801 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.933818 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.933843 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:00 crc kubenswrapper[4726]: I1004 03:42:00.933861 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:00Z","lastTransitionTime":"2025-10-04T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.036649 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.036720 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.036741 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.036765 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.036781 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.139398 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.139470 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.139494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.139522 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.139541 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.242176 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.242236 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.242324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.242358 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.242376 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.345616 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.345666 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.345682 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.345705 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.345721 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.448753 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.448840 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.448865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.448895 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.448917 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.502339 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.502391 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.502445 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:01 crc kubenswrapper[4726]: E1004 03:42:01.502693 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:01 crc kubenswrapper[4726]: E1004 03:42:01.502741 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:01 crc kubenswrapper[4726]: E1004 03:42:01.502809 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.552563 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.552612 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.552629 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.552652 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.552669 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.655521 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.655625 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.655645 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.655668 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.655684 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.759251 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.759310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.759332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.759363 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.759389 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.861903 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.862006 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.862029 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.862056 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.862076 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.965995 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.966151 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.966175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.966204 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:01 crc kubenswrapper[4726]: I1004 03:42:01.966224 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:01Z","lastTransitionTime":"2025-10-04T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.069135 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.069292 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.069319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.069347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.069369 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.172044 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.172148 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.172175 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.172205 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.172230 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.274653 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.275192 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.275381 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.275559 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.275872 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.379832 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.380245 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.380462 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.380624 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.380766 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.484827 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.485219 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.485441 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.485680 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.485868 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.502304 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:02 crc kubenswrapper[4726]: E1004 03:42:02.502813 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.503375 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:42:02 crc kubenswrapper[4726]: E1004 03:42:02.503642 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.588862 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.588916 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.588935 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.588960 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.588978 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.693226 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.693310 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.693328 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.693350 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.693367 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.796879 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.796954 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.796977 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.797009 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.797033 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.900621 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.900701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.900719 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.900742 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:02 crc kubenswrapper[4726]: I1004 03:42:02.900760 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:02Z","lastTransitionTime":"2025-10-04T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.002834 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.002865 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.002873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.002886 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.002895 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.105793 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.105859 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.105882 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.105915 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.105937 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.209174 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.209545 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.209564 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.209588 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.209605 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.312617 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.312679 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.312694 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.312717 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.312733 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.415511 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.415582 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.415604 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.415633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.415654 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.502220 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.502268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:03 crc kubenswrapper[4726]: E1004 03:42:03.502365 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:03 crc kubenswrapper[4726]: E1004 03:42:03.502463 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.502720 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:03 crc kubenswrapper[4726]: E1004 03:42:03.502936 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.518449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.518711 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.518871 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.519052 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.519278 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.622873 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.623267 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.623494 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.623523 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.623543 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.726598 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.726658 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.726707 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.726732 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.726752 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.829324 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.829354 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.829362 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.829374 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.829384 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.932208 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.932246 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.932256 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.932272 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:03 crc kubenswrapper[4726]: I1004 03:42:03.932282 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:03Z","lastTransitionTime":"2025-10-04T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.035633 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.035687 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.035698 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.035714 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.035722 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.138729 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.138788 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.138800 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.138863 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.138879 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.240829 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.240908 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.240926 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.240951 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.240968 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.343489 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.343547 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.343566 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.343589 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.343606 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.446816 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.446871 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.446888 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.446907 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.446920 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.501557 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:04 crc kubenswrapper[4726]: E1004 03:42:04.501757 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.550249 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.550285 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.550297 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.550309 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.550318 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.653261 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.653319 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.653332 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.653347 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.653357 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.756449 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.756509 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.756528 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.756552 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.756572 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.859654 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.859689 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.859701 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.859721 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.859734 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.961987 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.962045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.962057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.962074 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:04 crc kubenswrapper[4726]: I1004 03:42:04.962089 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:04Z","lastTransitionTime":"2025-10-04T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.065000 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.065045 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.065057 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.065072 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.065083 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:05Z","lastTransitionTime":"2025-10-04T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.084232 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.084274 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.084288 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.084305 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.084315 4726 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T03:42:05Z","lastTransitionTime":"2025-10-04T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.137954 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2"] Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.138699 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.141245 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.142527 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.143614 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.145179 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.176790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.176827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.176849 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.176868 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.177030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277738 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277774 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277810 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.277912 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.280523 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.290455 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.294076 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9671bc05-4b1d-4cf2-9a25-f4fafde7cac0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-gm7r2\" (UID: \"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.465048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.502260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:05 crc kubenswrapper[4726]: E1004 03:42:05.502457 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.502861 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:05 crc kubenswrapper[4726]: E1004 03:42:05.502994 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:05 crc kubenswrapper[4726]: I1004 03:42:05.503319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:05 crc kubenswrapper[4726]: E1004 03:42:05.503570 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:06 crc kubenswrapper[4726]: I1004 03:42:06.152997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" event={"ID":"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0","Type":"ContainerStarted","Data":"1579a556eb9f7bf688c605f7fc568e44aaeed2a3e24fd318e215a177bdc75661"} Oct 04 03:42:06 crc kubenswrapper[4726]: I1004 03:42:06.153075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" event={"ID":"9671bc05-4b1d-4cf2-9a25-f4fafde7cac0","Type":"ContainerStarted","Data":"fe6f1f299ee3b93d78fb7dfb6701af58a87880712e8c5522f46410b5f95dfe89"} Oct 04 03:42:06 crc kubenswrapper[4726]: I1004 03:42:06.173679 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-gm7r2" podStartSLOduration=78.173644319 podStartE2EDuration="1m18.173644319s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:06.172082678 +0000 UTC m=+100.346705931" watchObservedRunningTime="2025-10-04 03:42:06.173644319 +0000 UTC m=+100.348267602" Oct 04 03:42:06 crc kubenswrapper[4726]: I1004 03:42:06.502159 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:06 crc kubenswrapper[4726]: E1004 03:42:06.504171 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:06 crc kubenswrapper[4726]: I1004 03:42:06.795522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:06 crc kubenswrapper[4726]: E1004 03:42:06.795727 4726 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:42:06 crc kubenswrapper[4726]: E1004 03:42:06.795841 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs podName:77b969b7-0698-4589-a6cf-c08cc779ffe2 nodeName:}" failed. No retries permitted until 2025-10-04 03:43:10.795813472 +0000 UTC m=+164.970436715 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs") pod "network-metrics-daemon-9qn78" (UID: "77b969b7-0698-4589-a6cf-c08cc779ffe2") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 03:42:07 crc kubenswrapper[4726]: I1004 03:42:07.502256 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:07 crc kubenswrapper[4726]: E1004 03:42:07.502480 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:07 crc kubenswrapper[4726]: I1004 03:42:07.502540 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:07 crc kubenswrapper[4726]: I1004 03:42:07.502643 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:07 crc kubenswrapper[4726]: E1004 03:42:07.502703 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:07 crc kubenswrapper[4726]: E1004 03:42:07.503186 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:08 crc kubenswrapper[4726]: I1004 03:42:08.502096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:08 crc kubenswrapper[4726]: E1004 03:42:08.502243 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:09 crc kubenswrapper[4726]: I1004 03:42:09.501608 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:09 crc kubenswrapper[4726]: E1004 03:42:09.501990 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:09 crc kubenswrapper[4726]: I1004 03:42:09.501758 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:09 crc kubenswrapper[4726]: E1004 03:42:09.502052 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:09 crc kubenswrapper[4726]: I1004 03:42:09.501684 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:09 crc kubenswrapper[4726]: E1004 03:42:09.502242 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:10 crc kubenswrapper[4726]: I1004 03:42:10.502037 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:10 crc kubenswrapper[4726]: E1004 03:42:10.502277 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:11 crc kubenswrapper[4726]: I1004 03:42:11.501766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:11 crc kubenswrapper[4726]: I1004 03:42:11.501766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:11 crc kubenswrapper[4726]: E1004 03:42:11.502061 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:11 crc kubenswrapper[4726]: E1004 03:42:11.502219 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:11 crc kubenswrapper[4726]: I1004 03:42:11.501866 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:11 crc kubenswrapper[4726]: E1004 03:42:11.502407 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:12 crc kubenswrapper[4726]: I1004 03:42:12.502069 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:12 crc kubenswrapper[4726]: E1004 03:42:12.502437 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:13 crc kubenswrapper[4726]: I1004 03:42:13.502605 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:13 crc kubenswrapper[4726]: I1004 03:42:13.502739 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:13 crc kubenswrapper[4726]: I1004 03:42:13.502739 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:13 crc kubenswrapper[4726]: E1004 03:42:13.502885 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:13 crc kubenswrapper[4726]: E1004 03:42:13.503068 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:13 crc kubenswrapper[4726]: E1004 03:42:13.503274 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:14 crc kubenswrapper[4726]: I1004 03:42:14.502250 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:14 crc kubenswrapper[4726]: E1004 03:42:14.502446 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:14 crc kubenswrapper[4726]: I1004 03:42:14.503498 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:42:14 crc kubenswrapper[4726]: E1004 03:42:14.503776 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-snjd6_openshift-ovn-kubernetes(a8d6c450-3310-4058-90dc-0a39ab1934c4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" Oct 04 03:42:15 crc kubenswrapper[4726]: I1004 03:42:15.501737 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:15 crc kubenswrapper[4726]: E1004 03:42:15.501900 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:15 crc kubenswrapper[4726]: I1004 03:42:15.501923 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:15 crc kubenswrapper[4726]: I1004 03:42:15.501975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:15 crc kubenswrapper[4726]: E1004 03:42:15.502344 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:15 crc kubenswrapper[4726]: E1004 03:42:15.502465 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:16 crc kubenswrapper[4726]: I1004 03:42:16.501808 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:16 crc kubenswrapper[4726]: E1004 03:42:16.503615 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:17 crc kubenswrapper[4726]: I1004 03:42:17.501963 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:17 crc kubenswrapper[4726]: I1004 03:42:17.501979 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:17 crc kubenswrapper[4726]: E1004 03:42:17.502192 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:17 crc kubenswrapper[4726]: I1004 03:42:17.501985 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:17 crc kubenswrapper[4726]: E1004 03:42:17.502346 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:17 crc kubenswrapper[4726]: E1004 03:42:17.502415 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:18 crc kubenswrapper[4726]: I1004 03:42:18.501960 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:18 crc kubenswrapper[4726]: E1004 03:42:18.502177 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:19 crc kubenswrapper[4726]: I1004 03:42:19.501468 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:19 crc kubenswrapper[4726]: I1004 03:42:19.501480 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:19 crc kubenswrapper[4726]: I1004 03:42:19.501683 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:19 crc kubenswrapper[4726]: E1004 03:42:19.501678 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:19 crc kubenswrapper[4726]: E1004 03:42:19.501756 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:19 crc kubenswrapper[4726]: E1004 03:42:19.502047 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:20 crc kubenswrapper[4726]: I1004 03:42:20.501946 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:20 crc kubenswrapper[4726]: E1004 03:42:20.502156 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:21 crc kubenswrapper[4726]: I1004 03:42:21.501920 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:21 crc kubenswrapper[4726]: I1004 03:42:21.501974 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:21 crc kubenswrapper[4726]: I1004 03:42:21.502031 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:21 crc kubenswrapper[4726]: E1004 03:42:21.502291 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:21 crc kubenswrapper[4726]: E1004 03:42:21.502424 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:21 crc kubenswrapper[4726]: E1004 03:42:21.502553 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:22 crc kubenswrapper[4726]: I1004 03:42:22.502610 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:22 crc kubenswrapper[4726]: E1004 03:42:22.502855 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.212272 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/1.log" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.213226 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/0.log" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.213311 4726 generic.go:334] "Generic (PLEG): container finished" podID="7fa9344a-8c9c-4bc5-b91e-b6c1400d4025" containerID="f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4" exitCode=1 Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.213362 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerDied","Data":"f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4"} Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.213414 4726 scope.go:117] "RemoveContainer" containerID="6db2eb8420bdb424cce89bec2f62bc942659950300954553b04161c8ac348ccd" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.215143 4726 scope.go:117] "RemoveContainer" containerID="f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4" Oct 04 03:42:23 crc kubenswrapper[4726]: E1004 03:42:23.215538 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5c96v_openshift-multus(7fa9344a-8c9c-4bc5-b91e-b6c1400d4025)\"" pod="openshift-multus/multus-5c96v" podUID="7fa9344a-8c9c-4bc5-b91e-b6c1400d4025" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.501589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.501634 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:23 crc kubenswrapper[4726]: I1004 03:42:23.501640 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:23 crc kubenswrapper[4726]: E1004 03:42:23.501778 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:23 crc kubenswrapper[4726]: E1004 03:42:23.501887 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:23 crc kubenswrapper[4726]: E1004 03:42:23.501948 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:24 crc kubenswrapper[4726]: I1004 03:42:24.220075 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/1.log" Oct 04 03:42:24 crc kubenswrapper[4726]: I1004 03:42:24.502632 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:24 crc kubenswrapper[4726]: E1004 03:42:24.502843 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:25 crc kubenswrapper[4726]: I1004 03:42:25.501501 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:25 crc kubenswrapper[4726]: I1004 03:42:25.501551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:25 crc kubenswrapper[4726]: E1004 03:42:25.501628 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:25 crc kubenswrapper[4726]: E1004 03:42:25.501771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:25 crc kubenswrapper[4726]: I1004 03:42:25.502180 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:25 crc kubenswrapper[4726]: E1004 03:42:25.502387 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:26 crc kubenswrapper[4726]: I1004 03:42:26.501694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:26 crc kubenswrapper[4726]: E1004 03:42:26.503135 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:26 crc kubenswrapper[4726]: I1004 03:42:26.504798 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:42:26 crc kubenswrapper[4726]: E1004 03:42:26.534803 4726 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 03:42:26 crc kubenswrapper[4726]: E1004 03:42:26.643194 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.235853 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.241389 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.242913 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerStarted","Data":"66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188"} Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.245723 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.280581 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podStartSLOduration=99.280558358 podStartE2EDuration="1m39.280558358s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:27.27947299 +0000 UTC m=+121.454096223" watchObservedRunningTime="2025-10-04 03:42:27.280558358 +0000 UTC m=+121.455181611" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.475513 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9qn78"] Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.475716 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:27 crc kubenswrapper[4726]: E1004 03:42:27.475899 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.501965 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.502083 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:27 crc kubenswrapper[4726]: I1004 03:42:27.501987 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:27 crc kubenswrapper[4726]: E1004 03:42:27.502242 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:27 crc kubenswrapper[4726]: E1004 03:42:27.502405 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:27 crc kubenswrapper[4726]: E1004 03:42:27.502565 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:29 crc kubenswrapper[4726]: I1004 03:42:29.501793 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:29 crc kubenswrapper[4726]: I1004 03:42:29.501903 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:29 crc kubenswrapper[4726]: I1004 03:42:29.501831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:29 crc kubenswrapper[4726]: I1004 03:42:29.501821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:29 crc kubenswrapper[4726]: E1004 03:42:29.502092 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:29 crc kubenswrapper[4726]: E1004 03:42:29.502213 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:29 crc kubenswrapper[4726]: E1004 03:42:29.502333 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:29 crc kubenswrapper[4726]: E1004 03:42:29.502599 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:31 crc kubenswrapper[4726]: I1004 03:42:31.502009 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:31 crc kubenswrapper[4726]: I1004 03:42:31.502015 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:31 crc kubenswrapper[4726]: I1004 03:42:31.503389 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:31 crc kubenswrapper[4726]: E1004 03:42:31.503585 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:31 crc kubenswrapper[4726]: I1004 03:42:31.503646 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:31 crc kubenswrapper[4726]: E1004 03:42:31.503831 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:31 crc kubenswrapper[4726]: E1004 03:42:31.504017 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:31 crc kubenswrapper[4726]: E1004 03:42:31.504187 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:31 crc kubenswrapper[4726]: E1004 03:42:31.644768 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:42:33 crc kubenswrapper[4726]: I1004 03:42:33.501782 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:33 crc kubenswrapper[4726]: I1004 03:42:33.501835 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:33 crc kubenswrapper[4726]: I1004 03:42:33.501971 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:33 crc kubenswrapper[4726]: E1004 03:42:33.501989 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:33 crc kubenswrapper[4726]: I1004 03:42:33.502018 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:33 crc kubenswrapper[4726]: E1004 03:42:33.502178 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:33 crc kubenswrapper[4726]: E1004 03:42:33.502266 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:33 crc kubenswrapper[4726]: E1004 03:42:33.502390 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:35 crc kubenswrapper[4726]: I1004 03:42:35.501878 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:35 crc kubenswrapper[4726]: I1004 03:42:35.501916 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:35 crc kubenswrapper[4726]: I1004 03:42:35.501916 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:35 crc kubenswrapper[4726]: I1004 03:42:35.501980 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:35 crc kubenswrapper[4726]: E1004 03:42:35.502751 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:35 crc kubenswrapper[4726]: E1004 03:42:35.502855 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:35 crc kubenswrapper[4726]: E1004 03:42:35.502952 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:35 crc kubenswrapper[4726]: E1004 03:42:35.502634 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:36 crc kubenswrapper[4726]: I1004 03:42:36.503967 4726 scope.go:117] "RemoveContainer" containerID="f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4" Oct 04 03:42:36 crc kubenswrapper[4726]: E1004 03:42:36.645499 4726 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.284986 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/1.log" Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.285071 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerStarted","Data":"5db07f4304cf2eab69f21bcceb0014c80a2a876d97a3ce476e76b3e8dc8880d6"} Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.502082 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:37 crc kubenswrapper[4726]: E1004 03:42:37.502261 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.502096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:37 crc kubenswrapper[4726]: E1004 03:42:37.502367 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.502094 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:37 crc kubenswrapper[4726]: E1004 03:42:37.502471 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:37 crc kubenswrapper[4726]: I1004 03:42:37.502096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:37 crc kubenswrapper[4726]: E1004 03:42:37.502562 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:39 crc kubenswrapper[4726]: I1004 03:42:39.501866 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:39 crc kubenswrapper[4726]: E1004 03:42:39.502387 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:39 crc kubenswrapper[4726]: I1004 03:42:39.501911 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:39 crc kubenswrapper[4726]: E1004 03:42:39.502509 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:39 crc kubenswrapper[4726]: I1004 03:42:39.502000 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:39 crc kubenswrapper[4726]: E1004 03:42:39.502612 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:39 crc kubenswrapper[4726]: I1004 03:42:39.501866 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:39 crc kubenswrapper[4726]: E1004 03:42:39.502732 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:41 crc kubenswrapper[4726]: I1004 03:42:41.502048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:41 crc kubenswrapper[4726]: I1004 03:42:41.502051 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:41 crc kubenswrapper[4726]: E1004 03:42:41.502261 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 03:42:41 crc kubenswrapper[4726]: I1004 03:42:41.502079 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:41 crc kubenswrapper[4726]: E1004 03:42:41.502330 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 03:42:41 crc kubenswrapper[4726]: I1004 03:42:41.502078 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:41 crc kubenswrapper[4726]: E1004 03:42:41.502442 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 03:42:41 crc kubenswrapper[4726]: E1004 03:42:41.502549 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9qn78" podUID="77b969b7-0698-4589-a6cf-c08cc779ffe2" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.502324 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.502324 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.502459 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.503441 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.507568 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.507959 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.507998 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.508667 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.509350 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 03:42:43 crc kubenswrapper[4726]: I1004 03:42:43.509848 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.843981 4726 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.891939 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2s6pg"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.894346 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.897579 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.899001 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.902896 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.906352 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.909694 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.910419 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.910643 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.910903 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.911234 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.913384 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ffl97"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.913712 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.914061 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.932629 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.933061 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.934603 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.937276 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.939017 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.939288 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.939433 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.939687 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.941263 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.941418 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.942861 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cptzs"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.943087 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.943556 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.943710 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.943789 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.944169 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.944547 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.946624 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-image-import-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947445 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-node-pullsecrets\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947756 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhmwr\" (UniqueName: \"kubernetes.io/projected/d0883b34-7264-4f42-b041-c8bb3b742a3d-kube-api-access-hhmwr\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.947912 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948097 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-encryption-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948308 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948473 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948639 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948795 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit-dir\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-client\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949169 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949468 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glqrt\" (UniqueName: \"kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949627 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-serving-cert\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949908 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.948589 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.950781 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.951234 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.951724 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.949816 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.952752 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.952827 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.956573 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.956908 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.957091 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.957376 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ln2m"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.957815 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-nw2zq"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.958160 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.958313 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.958597 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.958677 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.958846 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.959487 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.961480 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.961834 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.962196 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963330 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963630 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963736 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963823 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963905 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.963974 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.964051 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.965490 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.966023 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.966999 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.967177 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.967317 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.967503 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.967693 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.967839 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.968176 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.972583 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.972946 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.973183 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.973803 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.973975 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.974159 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.974446 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.974792 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.975019 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.975295 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.975582 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.975786 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.975954 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.976297 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.976643 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.976904 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wp2cj"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.976910 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.976982 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.977335 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.978054 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.978480 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.978761 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.978974 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.979171 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.979239 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.980139 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.982870 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.982953 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.985163 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.985619 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg"] Oct 04 03:42:45 crc kubenswrapper[4726]: I1004 03:42:45.997631 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7ggk"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.010539 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.013688 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2277j"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.014243 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.014760 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.015266 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.020475 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.021302 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.021524 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.021860 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.022503 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zx9sv"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.023650 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.023978 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.027716 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.029199 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.030696 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.030880 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.031034 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.031202 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.031391 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.031532 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.031644 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.032017 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.032175 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.032273 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.032305 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.033456 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.033733 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.034055 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.034227 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.034358 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.034516 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.036970 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.037260 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.037357 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.037449 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.037918 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.037943 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.038667 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.039835 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.042161 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxrt5"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.042759 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.044778 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.045581 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.045912 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.046678 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.046880 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.047030 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.047314 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.048374 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-serving-cert\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050681 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1828b5f3-26e7-4b77-aefb-e25c713d7a37-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f284fb82-6924-4615-b695-af1d7de0ec31-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050900 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050923 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f284fb82-6924-4615-b695-af1d7de0ec31-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050950 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q86mv\" (UniqueName: \"kubernetes.io/projected/f284fb82-6924-4615-b695-af1d7de0ec31-kube-api-access-q86mv\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.050995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051035 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051066 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051087 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0444b9ed-94cf-4aab-84a7-f13208c48f64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051160 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff596\" (UniqueName: \"kubernetes.io/projected/1da19f66-1ae7-4e73-8f6c-84c471711441-kube-api-access-ff596\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-image-import-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051212 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2vr\" (UniqueName: \"kubernetes.io/projected/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-kube-api-access-7h2vr\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051237 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-auth-proxy-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-node-pullsecrets\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051274 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhmwr\" (UniqueName: \"kubernetes.io/projected/d0883b34-7264-4f42-b041-c8bb3b742a3d-kube-api-access-hhmwr\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051289 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6g7m\" (UniqueName: \"kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051322 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-encryption-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051338 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1da19f66-1ae7-4e73-8f6c-84c471711441-serving-cert\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051355 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051387 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88x79\" (UniqueName: \"kubernetes.io/projected/ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d-kube-api-access-88x79\") pod \"downloads-7954f5f757-nw2zq\" (UID: \"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d\") " pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051419 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1622c3b3-4111-4b64-9c2a-4b1a018ff536-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051434 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-config\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051452 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fzx\" (UniqueName: \"kubernetes.io/projected/0444b9ed-94cf-4aab-84a7-f13208c48f64-kube-api-access-q2fzx\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051468 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051483 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-client\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051519 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2fw8\" (UniqueName: \"kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051534 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-config\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051550 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b5c64dda-8b0b-404e-a875-037d3d7becb3-machine-approver-tls\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051565 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit-dir\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0444b9ed-94cf-4aab-84a7-f13208c48f64-serving-cert\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051614 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6t6s\" (UniqueName: \"kubernetes.io/projected/1622c3b3-4111-4b64-9c2a-4b1a018ff536-kube-api-access-v6t6s\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051648 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-trusted-ca\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-encryption-config\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051678 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-policies\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051693 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-dir\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051709 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1622c3b3-4111-4b64-9c2a-4b1a018ff536-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051725 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crpzq\" (UniqueName: \"kubernetes.io/projected/1828b5f3-26e7-4b77-aefb-e25c713d7a37-kube-api-access-crpzq\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051741 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-client\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051756 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7386c233-34f0-4eef-87d0-ac2b7b145897-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051771 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-serving-cert\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051787 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051802 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnppl\" (UniqueName: \"kubernetes.io/projected/819db0bd-0b6e-407f-bdde-3aaa513d66a2-kube-api-access-lnppl\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051828 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051845 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glqrt\" (UniqueName: \"kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051861 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26888\" (UniqueName: \"kubernetes.io/projected/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-kube-api-access-26888\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051878 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bn7x\" (UniqueName: \"kubernetes.io/projected/b5c64dda-8b0b-404e-a875-037d3d7becb3-kube-api-access-5bn7x\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051900 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051919 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-config\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051934 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq8j4\" (UniqueName: \"kubernetes.io/projected/7386c233-34f0-4eef-87d0-ac2b7b145897-kube-api-access-gq8j4\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1828b5f3-26e7-4b77-aefb-e25c713d7a37-proxy-tls\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051968 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.051985 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052001 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-images\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052016 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvv7\" (UniqueName: \"kubernetes.io/projected/1aad645d-c461-40b5-94e4-b5312f0a3ff7-kube-api-access-hbvv7\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb88v\" (UniqueName: \"kubernetes.io/projected/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-kube-api-access-tb88v\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052066 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/819db0bd-0b6e-407f-bdde-3aaa513d66a2-serving-cert\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052120 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.052212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit-dir\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.053408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.054090 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.054533 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.054959 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d0883b34-7264-4f42-b041-c8bb3b742a3d-node-pullsecrets\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.056445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.056524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.057635 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-serving-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.057854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-audit\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.057861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d0883b34-7264-4f42-b041-c8bb3b742a3d-image-import-ca\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.059011 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.059513 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.059812 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8mmdc"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.060926 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.061073 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.061463 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.061775 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.061605 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-etcd-client\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.062409 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.062938 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.063415 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.063678 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-encryption-config\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.063753 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.063914 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.066604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.074151 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.074291 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.074408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0883b34-7264-4f42-b041-c8bb3b742a3d-serving-cert\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.074460 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.077784 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.084211 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mhtxp"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.084711 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.084760 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.085525 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.085531 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.086631 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.087019 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.087218 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.093357 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.093967 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.094658 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h77cw"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.095044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.095568 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.095841 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.100494 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.101793 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.101973 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.105737 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ffl97"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.107861 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.119062 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.119881 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cptzs"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.121345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2s6pg"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.123129 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.123159 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ln2m"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.124412 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.125147 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.126384 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wp2cj"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.131874 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.136418 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-28vhf"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.137277 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.137370 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.138357 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.140494 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.143014 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8mmdc"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.144681 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.147417 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.147455 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.151403 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2277j"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153329 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q86mv\" (UniqueName: \"kubernetes.io/projected/f284fb82-6924-4615-b695-af1d7de0ec31-kube-api-access-q86mv\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153362 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0444b9ed-94cf-4aab-84a7-f13208c48f64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153436 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153453 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6g7m\" (UniqueName: \"kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1da19f66-1ae7-4e73-8f6c-84c471711441-serving-cert\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153648 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fzx\" (UniqueName: \"kubernetes.io/projected/0444b9ed-94cf-4aab-84a7-f13208c48f64-kube-api-access-q2fzx\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153680 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153708 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-client\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153776 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153799 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2fw8\" (UniqueName: \"kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-srv-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153899 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02db15c4-f787-4ed7-be15-bf9f7e59b342-metrics-tls\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b5c64dda-8b0b-404e-a875-037d3d7becb3-machine-approver-tls\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0444b9ed-94cf-4aab-84a7-f13208c48f64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.153942 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-config\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154020 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5bhk\" (UniqueName: \"kubernetes.io/projected/4c620617-c3fe-4350-aa40-869d930d9362-kube-api-access-j5bhk\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154049 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154074 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0444b9ed-94cf-4aab-84a7-f13208c48f64-serving-cert\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154126 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-policies\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154141 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-dir\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154157 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1622c3b3-4111-4b64-9c2a-4b1a018ff536-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crpzq\" (UniqueName: \"kubernetes.io/projected/1828b5f3-26e7-4b77-aefb-e25c713d7a37-kube-api-access-crpzq\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154193 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154211 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7386c233-34f0-4eef-87d0-ac2b7b145897-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154227 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154255 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154271 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d91c093-6638-48e4-aa66-6b096cb68cf9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154293 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26888\" (UniqueName: \"kubernetes.io/projected/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-kube-api-access-26888\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154309 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154371 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-config\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq8j4\" (UniqueName: \"kubernetes.io/projected/7386c233-34f0-4eef-87d0-ac2b7b145897-kube-api-access-gq8j4\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154459 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-images\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvv7\" (UniqueName: \"kubernetes.io/projected/1aad645d-c461-40b5-94e4-b5312f0a3ff7-kube-api-access-hbvv7\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154507 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffdtx\" (UniqueName: \"kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb88v\" (UniqueName: \"kubernetes.io/projected/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-kube-api-access-tb88v\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154583 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/819db0bd-0b6e-407f-bdde-3aaa513d66a2-serving-cert\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154608 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-key\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154651 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-config\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154666 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154704 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154731 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f284fb82-6924-4615-b695-af1d7de0ec31-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f284fb82-6924-4615-b695-af1d7de0ec31-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154770 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154780 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154818 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154841 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0024909c-8d97-45bc-bcd8-a2dba7a101cb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154863 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154881 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-serving-cert\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd7d8\" (UniqueName: \"kubernetes.io/projected/01bcb480-958d-4346-b185-44d717844903-kube-api-access-fd7d8\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154929 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154947 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff596\" (UniqueName: \"kubernetes.io/projected/1da19f66-1ae7-4e73-8f6c-84c471711441-kube-api-access-ff596\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.154966 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2vr\" (UniqueName: \"kubernetes.io/projected/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-kube-api-access-7h2vr\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155013 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-auth-proxy-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155028 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn7nx\" (UniqueName: \"kubernetes.io/projected/607637ee-8700-4690-9d42-0a7618d85c73-kube-api-access-vn7nx\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155044 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155073 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155089 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-cabundle\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155118 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155137 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88x79\" (UniqueName: \"kubernetes.io/projected/ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d-kube-api-access-88x79\") pod \"downloads-7954f5f757-nw2zq\" (UID: \"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d\") " pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155154 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1622c3b3-4111-4b64-9c2a-4b1a018ff536-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155175 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-config\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155190 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-etcd-client\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155210 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-apiservice-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155246 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6f6f\" (UniqueName: \"kubernetes.io/projected/654dedf3-d44a-4b22-bce8-6623f832a8f4-kube-api-access-v6f6f\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr267\" (UniqueName: \"kubernetes.io/projected/7b069e2c-070d-4051-85b6-3af46fb2e22c-kube-api-access-wr267\") pod \"migrator-59844c95c7-7ksbd\" (UID: \"7b069e2c-070d-4051-85b6-3af46fb2e22c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-config\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155318 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlh8k\" (UniqueName: \"kubernetes.io/projected/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-kube-api-access-mlh8k\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155337 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155375 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2nkx\" (UniqueName: \"kubernetes.io/projected/02db15c4-f787-4ed7-be15-bf9f7e59b342-kube-api-access-q2nkx\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155411 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-encryption-config\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6t6s\" (UniqueName: \"kubernetes.io/projected/1622c3b3-4111-4b64-9c2a-4b1a018ff536-kube-api-access-v6t6s\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155443 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-trusted-ca\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-serving-cert\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnppl\" (UniqueName: \"kubernetes.io/projected/819db0bd-0b6e-407f-bdde-3aaa513d66a2-kube-api-access-lnppl\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqftk\" (UniqueName: \"kubernetes.io/projected/986c01ee-6bf6-4654-a8b4-96ee71ace277-kube-api-access-bqftk\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bn7x\" (UniqueName: \"kubernetes.io/projected/b5c64dda-8b0b-404e-a875-037d3d7becb3-kube-api-access-5bn7x\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155574 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01bcb480-958d-4346-b185-44d717844903-tmpfs\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155579 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155592 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1828b5f3-26e7-4b77-aefb-e25c713d7a37-proxy-tls\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155640 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155656 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qq46\" (UniqueName: \"kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155692 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155707 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155725 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155743 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1828b5f3-26e7-4b77-aefb-e25c713d7a37-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155759 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-webhook-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.155791 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-service-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.156225 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-config\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.158645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-policies\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.158699 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1aad645d-c461-40b5-94e4-b5312f0a3ff7-audit-dir\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.158826 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1da19f66-1ae7-4e73-8f6c-84c471711441-serving-cert\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.159295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.159544 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.159652 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.159992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c64dda-8b0b-404e-a875-037d3d7becb3-auth-proxy-config\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160313 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160396 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/b5c64dda-8b0b-404e-a875-037d3d7becb3-machine-approver-tls\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160643 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1622c3b3-4111-4b64-9c2a-4b1a018ff536-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160715 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160784 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7ggk"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.159861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7386c233-34f0-4eef-87d0-ac2b7b145897-images\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.160873 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.161352 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1aad645d-c461-40b5-94e4-b5312f0a3ff7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.161572 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-etcd-client\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.161801 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-config\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.162036 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f284fb82-6924-4615-b695-af1d7de0ec31-config\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.162473 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1622c3b3-4111-4b64-9c2a-4b1a018ff536-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.162912 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.163001 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/819db0bd-0b6e-407f-bdde-3aaa513d66a2-serving-cert\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.163433 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.163776 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1da19f66-1ae7-4e73-8f6c-84c471711441-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.163964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7386c233-34f0-4eef-87d0-ac2b7b145897-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.164423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.164442 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/819db0bd-0b6e-407f-bdde-3aaa513d66a2-trusted-ca\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.164724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-serving-cert\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.164731 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1828b5f3-26e7-4b77-aefb-e25c713d7a37-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.165197 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1828b5f3-26e7-4b77-aefb-e25c713d7a37-proxy-tls\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.165235 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mhtxp"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.165783 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.166203 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0444b9ed-94cf-4aab-84a7-f13208c48f64-serving-cert\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.166198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1aad645d-c461-40b5-94e4-b5312f0a3ff7-encryption-config\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.167264 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxrt5"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.167415 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f284fb82-6924-4615-b695-af1d7de0ec31-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.168589 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.170860 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.170957 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.171027 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nw2zq"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.174385 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.174503 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.174585 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.176834 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.178623 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.180164 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.180215 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.182752 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l5gq"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.183970 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6vzk5"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.184077 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.184524 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.184603 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.184802 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l5gq"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.185805 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.186956 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-28vhf"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.187996 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h77cw"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.188992 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.190314 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.191321 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nqmms"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.194959 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.196991 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.199091 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nqmms"] Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.217364 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.238236 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256558 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256796 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256856 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256911 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.256985 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257022 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257063 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-srv-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257094 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257154 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02db15c4-f787-4ed7-be15-bf9f7e59b342-metrics-tls\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5bhk\" (UniqueName: \"kubernetes.io/projected/4c620617-c3fe-4350-aa40-869d930d9362-kube-api-access-j5bhk\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257328 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257405 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257465 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d91c093-6638-48e4-aa66-6b096cb68cf9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257634 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffdtx\" (UniqueName: \"kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257676 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-key\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257721 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257762 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.257799 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0024909c-8d97-45bc-bcd8-a2dba7a101cb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-serving-cert\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258565 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd7d8\" (UniqueName: \"kubernetes.io/projected/01bcb480-958d-4346-b185-44d717844903-kube-api-access-fd7d8\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258617 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn7nx\" (UniqueName: \"kubernetes.io/projected/607637ee-8700-4690-9d42-0a7618d85c73-kube-api-access-vn7nx\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258772 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-cabundle\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258850 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-etcd-client\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258886 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-apiservice-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258908 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.258986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6f6f\" (UniqueName: \"kubernetes.io/projected/654dedf3-d44a-4b22-bce8-6623f832a8f4-kube-api-access-v6f6f\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr267\" (UniqueName: \"kubernetes.io/projected/7b069e2c-070d-4051-85b6-3af46fb2e22c-kube-api-access-wr267\") pod \"migrator-59844c95c7-7ksbd\" (UID: \"7b069e2c-070d-4051-85b6-3af46fb2e22c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259159 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlh8k\" (UniqueName: \"kubernetes.io/projected/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-kube-api-access-mlh8k\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259185 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259285 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-config\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259310 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2nkx\" (UniqueName: \"kubernetes.io/projected/02db15c4-f787-4ed7-be15-bf9f7e59b342-kube-api-access-q2nkx\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqftk\" (UniqueName: \"kubernetes.io/projected/986c01ee-6bf6-4654-a8b4-96ee71ace277-kube-api-access-bqftk\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259531 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01bcb480-958d-4346-b185-44d717844903-tmpfs\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qq46\" (UniqueName: \"kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.259973 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01bcb480-958d-4346-b185-44d717844903-tmpfs\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260406 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260679 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260724 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260742 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-webhook-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-srv-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.260830 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-service-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.261796 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/607637ee-8700-4690-9d42-0a7618d85c73-profile-collector-cert\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.280245 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.296292 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.316201 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.320626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02db15c4-f787-4ed7-be15-bf9f7e59b342-metrics-tls\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.336708 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.357327 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.384228 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.412971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhmwr\" (UniqueName: \"kubernetes.io/projected/d0883b34-7264-4f42-b041-c8bb3b742a3d-kube-api-access-hhmwr\") pod \"apiserver-76f77b778f-2s6pg\" (UID: \"d0883b34-7264-4f42-b041-c8bb3b742a3d\") " pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.437549 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.437592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glqrt\" (UniqueName: \"kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt\") pod \"controller-manager-879f6c89f-pbtx9\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.458289 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.476739 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.498313 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.518191 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.536933 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.556620 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.557698 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.574302 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.576324 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.597346 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.618008 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.638284 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.653314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-serving-cert\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.657484 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.678564 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.680665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-config\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.685924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c620617-c3fe-4350-aa40-869d930d9362-etcd-client\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.697464 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.698493 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.719127 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.722632 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c620617-c3fe-4350-aa40-869d930d9362-etcd-service-ca\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.736536 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.757409 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.765757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-webhook-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.766355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01bcb480-958d-4346-b185-44d717844903-apiservice-cert\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.777394 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.792172 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:42:46 crc kubenswrapper[4726]: W1004 03:42:46.800673 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0854df0d_1ca6_41a4_ba86_98fdec338fc1.slice/crio-d3e0713ba5255ff56017c441115a57eb04ff28ef8b611a8ca8c7aa4bd7ee0ea0 WatchSource:0}: Error finding container d3e0713ba5255ff56017c441115a57eb04ff28ef8b611a8ca8c7aa4bd7ee0ea0: Status 404 returned error can't find the container with id d3e0713ba5255ff56017c441115a57eb04ff28ef8b611a8ca8c7aa4bd7ee0ea0 Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.806492 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.813204 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.817067 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.822004 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.828716 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2s6pg"] Oct 04 03:42:46 crc kubenswrapper[4726]: W1004 03:42:46.836463 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0883b34_7264_4f42_b041_c8bb3b742a3d.slice/crio-94f4ac9cc5e85e2abc2f37201722135a0235c57cdcced2d094295fe8ab1099fe WatchSource:0}: Error finding container 94f4ac9cc5e85e2abc2f37201722135a0235c57cdcced2d094295fe8ab1099fe: Status 404 returned error can't find the container with id 94f4ac9cc5e85e2abc2f37201722135a0235c57cdcced2d094295fe8ab1099fe Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.837884 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.879488 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.886334 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.894287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.894844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.896067 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.901221 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.917399 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.921728 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.937790 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.942636 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.958083 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.972846 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.976899 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 03:42:46 crc kubenswrapper[4726]: I1004 03:42:46.996922 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.017349 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.019230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.038222 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.045086 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.057683 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.059157 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.076827 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.094983 4726 request.go:700] Waited for 1.014146647s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/configmaps?fieldSelector=metadata.name%3Dv4-0-config-system-service-ca&limit=500&resourceVersion=0 Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.096367 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.100045 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.117721 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.137236 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.138728 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.156637 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.177380 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.187509 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.197908 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.216968 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.220775 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-cabundle\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.238636 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.255788 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/986c01ee-6bf6-4654-a8b4-96ee71ace277-signing-key\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.257646 4726 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.257676 4726 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.257769 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config podName:654dedf3-d44a-4b22-bce8-6623f832a8f4 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.757707663 +0000 UTC m=+141.932330886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config") pod "service-ca-operator-777779d784-h77cw" (UID: "654dedf3-d44a-4b22-bce8-6623f832a8f4") : failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.257796 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config podName:9d91c093-6638-48e4-aa66-6b096cb68cf9 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.757784505 +0000 UTC m=+141.932407728 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config") pod "kube-controller-manager-operator-78b949d7b-58k7b" (UID: "9d91c093-6638-48e4-aa66-6b096cb68cf9") : failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.258054 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258345 4726 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258509 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca podName:b6cc271a-fe0c-40ca-a28d-68568dcdf2d2 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.758468104 +0000 UTC m=+141.933091357 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca") pod "marketplace-operator-79b997595-t8p4b" (UID: "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2") : failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258544 4726 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258633 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert podName:0024909c-8d97-45bc-bcd8-a2dba7a101cb nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.758606518 +0000 UTC m=+141.933229781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert") pod "kube-apiserver-operator-766d6c64bb-ctsv5" (UID: "0024909c-8d97-45bc-bcd8-a2dba7a101cb") : failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258679 4726 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.258734 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert podName:9d91c093-6638-48e4-aa66-6b096cb68cf9 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.758721151 +0000 UTC m=+141.933344404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert") pod "kube-controller-manager-operator-78b949d7b-58k7b" (UID: "9d91c093-6638-48e4-aa66-6b096cb68cf9") : failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.259035 4726 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.259197 4726 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.260061 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config podName:0024909c-8d97-45bc-bcd8-a2dba7a101cb nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.760040418 +0000 UTC m=+141.934663671 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config") pod "kube-apiserver-operator-766d6c64bb-ctsv5" (UID: "0024909c-8d97-45bc-bcd8-a2dba7a101cb") : failed to sync configmap cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.260134 4726 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.260562 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert podName:654dedf3-d44a-4b22-bce8-6623f832a8f4 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.760528321 +0000 UTC m=+141.935151574 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert") pod "service-ca-operator-777779d784-h77cw" (UID: "654dedf3-d44a-4b22-bce8-6623f832a8f4") : failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: E1004 03:42:47.260604 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics podName:b6cc271a-fe0c-40ca-a28d-68568dcdf2d2 nodeName:}" failed. No retries permitted until 2025-10-04 03:42:47.760587643 +0000 UTC m=+141.935210966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics") pod "marketplace-operator-79b997595-t8p4b" (UID: "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2") : failed to sync secret cache: timed out waiting for the condition Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.277167 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.297027 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.317076 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.323996 4726 generic.go:334] "Generic (PLEG): container finished" podID="d0883b34-7264-4f42-b041-c8bb3b742a3d" containerID="8f151b87218b1ee42618a4a35f4db62abc2e68f86b4be39edd84a2a5b753d4b8" exitCode=0 Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.324147 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" event={"ID":"d0883b34-7264-4f42-b041-c8bb3b742a3d","Type":"ContainerDied","Data":"8f151b87218b1ee42618a4a35f4db62abc2e68f86b4be39edd84a2a5b753d4b8"} Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.324208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" event={"ID":"d0883b34-7264-4f42-b041-c8bb3b742a3d","Type":"ContainerStarted","Data":"94f4ac9cc5e85e2abc2f37201722135a0235c57cdcced2d094295fe8ab1099fe"} Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.325860 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" event={"ID":"0854df0d-1ca6-41a4-ba86-98fdec338fc1","Type":"ContainerStarted","Data":"d90f0bb0450edf32ee4ebcc8370e81d6b84509f0f507f4bec08d62d2b3c8eff7"} Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.325905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" event={"ID":"0854df0d-1ca6-41a4-ba86-98fdec338fc1","Type":"ContainerStarted","Data":"d3e0713ba5255ff56017c441115a57eb04ff28ef8b611a8ca8c7aa4bd7ee0ea0"} Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.337898 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.357347 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.377618 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.396324 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.417287 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.437215 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.457026 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.477803 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.505450 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.516839 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.537677 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.557303 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.577215 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.597875 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.616933 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.637368 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.656546 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.677082 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.697535 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.717187 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.758418 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.778163 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.794222 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.795759 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.795406 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d91c093-6638-48e4-aa66-6b096cb68cf9-config\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.795897 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.795979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.796092 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.796189 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.796285 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.796437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.797532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/654dedf3-d44a-4b22-bce8-6623f832a8f4-config\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.797646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.798042 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.799682 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0024909c-8d97-45bc-bcd8-a2dba7a101cb-config\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.803190 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/654dedf3-d44a-4b22-bce8-6623f832a8f4-serving-cert\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.804646 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d91c093-6638-48e4-aa66-6b096cb68cf9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.804784 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.806379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0024909c-8d97-45bc-bcd8-a2dba7a101cb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.857487 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q86mv\" (UniqueName: \"kubernetes.io/projected/f284fb82-6924-4615-b695-af1d7de0ec31-kube-api-access-q86mv\") pod \"openshift-apiserver-operator-796bbdcf4f-t6rc4\" (UID: \"f284fb82-6924-4615-b695-af1d7de0ec31\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.884914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fzx\" (UniqueName: \"kubernetes.io/projected/0444b9ed-94cf-4aab-84a7-f13208c48f64-kube-api-access-q2fzx\") pod \"openshift-config-operator-7777fb866f-cptzs\" (UID: \"0444b9ed-94cf-4aab-84a7-f13208c48f64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.899383 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq8j4\" (UniqueName: \"kubernetes.io/projected/7386c233-34f0-4eef-87d0-ac2b7b145897-kube-api-access-gq8j4\") pod \"machine-api-operator-5694c8668f-ffl97\" (UID: \"7386c233-34f0-4eef-87d0-ac2b7b145897\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.923748 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2fw8\" (UniqueName: \"kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8\") pod \"console-f9d7485db-x6s4v\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.940396 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.946857 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88x79\" (UniqueName: \"kubernetes.io/projected/ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d-kube-api-access-88x79\") pod \"downloads-7954f5f757-nw2zq\" (UID: \"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d\") " pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.961639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff596\" (UniqueName: \"kubernetes.io/projected/1da19f66-1ae7-4e73-8f6c-84c471711441-kube-api-access-ff596\") pod \"authentication-operator-69f744f599-2ln2m\" (UID: \"1da19f66-1ae7-4e73-8f6c-84c471711441\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.985554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2vr\" (UniqueName: \"kubernetes.io/projected/4c8fbfde-a722-4762-b6f4-b714a96f1c9f-kube-api-access-7h2vr\") pod \"multus-admission-controller-857f4d67dd-s7ggk\" (UID: \"4c8fbfde-a722-4762-b6f4-b714a96f1c9f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:47 crc kubenswrapper[4726]: I1004 03:42:47.997232 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvv7\" (UniqueName: \"kubernetes.io/projected/1aad645d-c461-40b5-94e4-b5312f0a3ff7-kube-api-access-hbvv7\") pod \"apiserver-7bbb656c7d-f7l7q\" (UID: \"1aad645d-c461-40b5-94e4-b5312f0a3ff7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.015552 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.017745 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb88v\" (UniqueName: \"kubernetes.io/projected/fa73eb1e-f78b-41f1-9be1-aeae979d66f9-kube-api-access-tb88v\") pod \"control-plane-machine-set-operator-78cbb6b69f-qstzg\" (UID: \"fa73eb1e-f78b-41f1-9be1-aeae979d66f9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.049069 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crpzq\" (UniqueName: \"kubernetes.io/projected/1828b5f3-26e7-4b77-aefb-e25c713d7a37-kube-api-access-crpzq\") pod \"machine-config-controller-84d6567774-2277j\" (UID: \"1828b5f3-26e7-4b77-aefb-e25c713d7a37\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.063483 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26888\" (UniqueName: \"kubernetes.io/projected/c09a4bf0-3a5c-44ff-a5fb-1f81188362e9-kube-api-access-26888\") pod \"cluster-samples-operator-665b6dd947-s6fzl\" (UID: \"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.082954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnppl\" (UniqueName: \"kubernetes.io/projected/819db0bd-0b6e-407f-bdde-3aaa513d66a2-kube-api-access-lnppl\") pod \"console-operator-58897d9998-wp2cj\" (UID: \"819db0bd-0b6e-407f-bdde-3aaa513d66a2\") " pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.086151 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.095745 4726 request.go:700] Waited for 1.933122034s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.104441 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.125837 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bn7x\" (UniqueName: \"kubernetes.io/projected/b5c64dda-8b0b-404e-a875-037d3d7becb3-kube-api-access-5bn7x\") pod \"machine-approver-56656f9798-gksb9\" (UID: \"b5c64dda-8b0b-404e-a875-037d3d7becb3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.132837 4726 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.132864 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.145134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6g7m\" (UniqueName: \"kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m\") pod \"route-controller-manager-6576b87f9c-ttws2\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.152629 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.154698 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.174244 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.181807 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5c64dda_8b0b_404e_a875_037d3d7becb3.slice/crio-d7c5aee50ab545143643fe03ff7a9efcd9bdcc3fad24c2995bd5ccca23e09aa2 WatchSource:0}: Error finding container d7c5aee50ab545143643fe03ff7a9efcd9bdcc3fad24c2995bd5ccca23e09aa2: Status 404 returned error can't find the container with id d7c5aee50ab545143643fe03ff7a9efcd9bdcc3fad24c2995bd5ccca23e09aa2 Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.193012 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.201948 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.211153 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.215127 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.221764 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.223835 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.230502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.238832 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.248651 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.253339 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.253894 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf284fb82_6924_4615_b695_af1d7de0ec31.slice/crio-1dcb71c6b9aa214d480c3d673a685779a8870f99bc27e93b00825932cd89d645 WatchSource:0}: Error finding container 1dcb71c6b9aa214d480c3d673a685779a8870f99bc27e93b00825932cd89d645: Status 404 returned error can't find the container with id 1dcb71c6b9aa214d480c3d673a685779a8870f99bc27e93b00825932cd89d645 Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.258684 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.271809 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.280139 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-s7ggk"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.281284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6t6s\" (UniqueName: \"kubernetes.io/projected/1622c3b3-4111-4b64-9c2a-4b1a018ff536-kube-api-access-v6t6s\") pod \"openshift-controller-manager-operator-756b6f6bc6-xrfvj\" (UID: \"1622c3b3-4111-4b64-9c2a-4b1a018ff536\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.292430 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.302017 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.308648 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.315729 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.333896 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" event={"ID":"f284fb82-6924-4615-b695-af1d7de0ec31","Type":"ContainerStarted","Data":"1dcb71c6b9aa214d480c3d673a685779a8870f99bc27e93b00825932cd89d645"} Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.335891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" event={"ID":"b5c64dda-8b0b-404e-a875-037d3d7becb3","Type":"ContainerStarted","Data":"d7c5aee50ab545143643fe03ff7a9efcd9bdcc3fad24c2995bd5ccca23e09aa2"} Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.336461 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.338073 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pbtx9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.338128 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.348762 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d91c093-6638-48e4-aa66-6b096cb68cf9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-58k7b\" (UID: \"9d91c093-6638-48e4-aa66-6b096cb68cf9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.358255 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ffl97"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.366150 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffdtx\" (UniqueName: \"kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx\") pod \"oauth-openshift-558db77b4-wssnh\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.387268 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5bhk\" (UniqueName: \"kubernetes.io/projected/4c620617-c3fe-4350-aa40-869d930d9362-kube-api-access-j5bhk\") pod \"etcd-operator-b45778765-8mmdc\" (UID: \"4c620617-c3fe-4350-aa40-869d930d9362\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.409608 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0024909c-8d97-45bc-bcd8-a2dba7a101cb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ctsv5\" (UID: \"0024909c-8d97-45bc-bcd8-a2dba7a101cb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.448210 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.451999 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.452689 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn7nx\" (UniqueName: \"kubernetes.io/projected/607637ee-8700-4690-9d42-0a7618d85c73-kube-api-access-vn7nx\") pod \"olm-operator-6b444d44fb-kcv7c\" (UID: \"607637ee-8700-4690-9d42-0a7618d85c73\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.467591 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6f6f\" (UniqueName: \"kubernetes.io/projected/654dedf3-d44a-4b22-bce8-6623f832a8f4-kube-api-access-v6f6f\") pod \"service-ca-operator-777779d784-h77cw\" (UID: \"654dedf3-d44a-4b22-bce8-6623f832a8f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.472565 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.479467 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.487754 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr267\" (UniqueName: \"kubernetes.io/projected/7b069e2c-070d-4051-85b6-3af46fb2e22c-kube-api-access-wr267\") pod \"migrator-59844c95c7-7ksbd\" (UID: \"7b069e2c-070d-4051-85b6-3af46fb2e22c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.491566 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.507984 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlh8k\" (UniqueName: \"kubernetes.io/projected/eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00-kube-api-access-mlh8k\") pod \"kube-storage-version-migrator-operator-b67b599dd-stsdb\" (UID: \"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.524124 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nw2zq"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.526276 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.526891 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2nkx\" (UniqueName: \"kubernetes.io/projected/02db15c4-f787-4ed7-be15-bf9f7e59b342-kube-api-access-q2nkx\") pod \"dns-operator-744455d44c-xxrt5\" (UID: \"02db15c4-f787-4ed7-be15-bf9f7e59b342\") " pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.542463 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.546778 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqftk\" (UniqueName: \"kubernetes.io/projected/986c01ee-6bf6-4654-a8b4-96ee71ace277-kube-api-access-bqftk\") pod \"service-ca-9c57cc56f-mhtxp\" (UID: \"986c01ee-6bf6-4654-a8b4-96ee71ace277\") " pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.557484 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.566852 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qq46\" (UniqueName: \"kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46\") pod \"marketplace-operator-79b997595-t8p4b\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.568217 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.598677 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cptzs"] Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.608983 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609062 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7jth\" (UniqueName: \"kubernetes.io/projected/42d8eec6-1736-44e1-adfd-995ecbde57a5-kube-api-access-w7jth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609222 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-images\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxblr\" (UniqueName: \"kubernetes.io/projected/3a92e571-ee54-4832-a6f5-c9409da8b83d-kube-api-access-bxblr\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-default-certificate\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609329 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609405 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609467 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a92e571-ee54-4832-a6f5-c9409da8b83d-proxy-tls\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609490 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609518 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj6c6\" (UniqueName: \"kubernetes.io/projected/503a1e39-10e4-4d8e-946a-410e83081176-kube-api-access-lj6c6\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609543 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609606 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609668 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jns9x\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-kube-api-access-jns9x\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609701 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609728 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-stats-auth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609749 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-srv-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609845 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.609952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610035 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610067 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ddfe3ae-32c0-4636-8a34-87b418aa209a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610146 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ddfe3ae-32c0-4636-8a34-87b418aa209a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610197 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-metrics-certs\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.610247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: E1004 03:42:48.610842 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.110825748 +0000 UTC m=+143.285448981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.611224 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.611890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/503a1e39-10e4-4d8e-946a-410e83081176-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.612053 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd2zj\" (UniqueName: \"kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.612145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613185 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsq4j\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613267 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42d8eec6-1736-44e1-adfd-995ecbde57a5-service-ca-bundle\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613293 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht526\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-kube-api-access-ht526\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613317 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lzpf\" (UniqueName: \"kubernetes.io/projected/2950e709-3e91-437c-8f7e-0c85fc073579-kube-api-access-5lzpf\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613342 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.613374 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.651312 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.661694 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714589 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714764 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714791 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ddfe3ae-32c0-4636-8a34-87b418aa209a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714843 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-metrics-certs\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714913 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714937 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-certs\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.714956 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-csi-data-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715002 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/503a1e39-10e4-4d8e-946a-410e83081176-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715021 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-mountpoint-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715046 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-node-bootstrap-token\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715121 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd2zj\" (UniqueName: \"kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715150 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715184 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c216c9a6-d84c-42f6-b900-0f0c37c1e107-config-volume\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ms7\" (UniqueName: \"kubernetes.io/projected/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-kube-api-access-p7ms7\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsq4j\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715285 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-cert\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715314 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42d8eec6-1736-44e1-adfd-995ecbde57a5-service-ca-bundle\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715334 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht526\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-kube-api-access-ht526\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715363 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lzpf\" (UniqueName: \"kubernetes.io/projected/2950e709-3e91-437c-8f7e-0c85fc073579-kube-api-access-5lzpf\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715383 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715400 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28ngc\" (UniqueName: \"kubernetes.io/projected/3617abd0-173f-4e18-a1de-8059039e542c-kube-api-access-28ngc\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715513 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7jth\" (UniqueName: \"kubernetes.io/projected/42d8eec6-1736-44e1-adfd-995ecbde57a5-kube-api-access-w7jth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715577 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-images\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715595 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8gtj\" (UniqueName: \"kubernetes.io/projected/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-kube-api-access-x8gtj\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715613 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715630 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxblr\" (UniqueName: \"kubernetes.io/projected/3a92e571-ee54-4832-a6f5-c9409da8b83d-kube-api-access-bxblr\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715656 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-default-certificate\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715770 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c216c9a6-d84c-42f6-b900-0f0c37c1e107-metrics-tls\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715809 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715846 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a92e571-ee54-4832-a6f5-c9409da8b83d-proxy-tls\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715872 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj6c6\" (UniqueName: \"kubernetes.io/projected/503a1e39-10e4-4d8e-946a-410e83081176-kube-api-access-lj6c6\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715906 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-plugins-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715924 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-registration-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.715965 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg8vk\" (UniqueName: \"kubernetes.io/projected/c216c9a6-d84c-42f6-b900-0f0c37c1e107-kube-api-access-kg8vk\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716004 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716022 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jns9x\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-kube-api-access-jns9x\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-stats-auth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716090 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-srv-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-socket-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716201 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ddfe3ae-32c0-4636-8a34-87b418aa209a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.716748 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: E1004 03:42:48.716863 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.216838958 +0000 UTC m=+143.391462191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.718388 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-images\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.719127 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.719722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3a92e571-ee54-4832-a6f5-c9409da8b83d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.719874 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.720292 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-default-certificate\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.720593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.720953 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42d8eec6-1736-44e1-adfd-995ecbde57a5-service-ca-bundle\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.722822 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ddfe3ae-32c0-4636-8a34-87b418aa209a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.723202 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.723882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.728524 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.728626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-metrics-certs\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.729448 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ddfe3ae-32c0-4636-8a34-87b418aa209a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.729632 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.730256 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.735417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3a92e571-ee54-4832-a6f5-c9409da8b83d-proxy-tls\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.737379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.765937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxblr\" (UniqueName: \"kubernetes.io/projected/3a92e571-ee54-4832-a6f5-c9409da8b83d-kube-api-access-bxblr\") pod \"machine-config-operator-74547568cd-xffj7\" (UID: \"3a92e571-ee54-4832-a6f5-c9409da8b83d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.772040 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.790453 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.794997 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jns9x\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-kube-api-access-jns9x\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.801018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.801700 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.802223 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-profile-collector-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.802633 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2950e709-3e91-437c-8f7e-0c85fc073579-srv-cert\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.802969 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.803571 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/42d8eec6-1736-44e1-adfd-995ecbde57a5-stats-auth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.805002 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c8fbfde_a722_4762_b6f4_b714a96f1c9f.slice/crio-a86a8429650cb36795ebed359f38580bc554c57a14187b41510f27bdbc819dfa WatchSource:0}: Error finding container a86a8429650cb36795ebed359f38580bc554c57a14187b41510f27bdbc819dfa: Status 404 returned error can't find the container with id a86a8429650cb36795ebed359f38580bc554c57a14187b41510f27bdbc819dfa Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.805705 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd7d8\" (UniqueName: \"kubernetes.io/projected/01bcb480-958d-4346-b185-44d717844903-kube-api-access-fd7d8\") pod \"packageserver-d55dfcdfc-w6fz4\" (UID: \"01bcb480-958d-4346-b185-44d717844903\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.808047 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7ddfe3ae-32c0-4636-8a34-87b418aa209a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tspxk\" (UID: \"7ddfe3ae-32c0-4636-8a34-87b418aa209a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.808898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/503a1e39-10e4-4d8e-946a-410e83081176-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817330 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c216c9a6-d84c-42f6-b900-0f0c37c1e107-metrics-tls\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817429 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-plugins-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817519 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-registration-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg8vk\" (UniqueName: \"kubernetes.io/projected/c216c9a6-d84c-42f6-b900-0f0c37c1e107-kube-api-access-kg8vk\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817650 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-socket-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817730 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-certs\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817857 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-csi-data-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817925 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-mountpoint-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.817983 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-node-bootstrap-token\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818052 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c216c9a6-d84c-42f6-b900-0f0c37c1e107-config-volume\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818099 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ms7\" (UniqueName: \"kubernetes.io/projected/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-kube-api-access-p7ms7\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818217 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-cert\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28ngc\" (UniqueName: \"kubernetes.io/projected/3617abd0-173f-4e18-a1de-8059039e542c-kube-api-access-28ngc\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-registration-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8gtj\" (UniqueName: \"kubernetes.io/projected/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-kube-api-access-x8gtj\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: E1004 03:42:48.818587 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.31856617 +0000 UTC m=+143.493189393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.818757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-plugins-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.819224 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-socket-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.819510 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-mountpoint-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.819621 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-csi-data-dir\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.821749 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c216c9a6-d84c-42f6-b900-0f0c37c1e107-config-volume\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.823477 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded1d5fb4_ef43_4fca_9a16_4f67b8537c0d.slice/crio-84448593f4575518589387ff7eae34e9d5287bc0fdf0f00d2896a642f4f6a44c WatchSource:0}: Error finding container 84448593f4575518589387ff7eae34e9d5287bc0fdf0f00d2896a642f4f6a44c: Status 404 returned error can't find the container with id 84448593f4575518589387ff7eae34e9d5287bc0fdf0f00d2896a642f4f6a44c Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.823938 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-certs\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.824390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3617abd0-173f-4e18-a1de-8059039e542c-node-bootstrap-token\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.826170 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-cert\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.827090 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd2zj\" (UniqueName: \"kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj\") pod \"collect-profiles-29325810-gkptq\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.829396 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2ca287e_027c_44b2_8068_b84437de7e71.slice/crio-d4f66349559f4a30ec7a570ba461830dcd2cde794a40a1d6f26a9657e10412bb WatchSource:0}: Error finding container d4f66349559f4a30ec7a570ba461830dcd2cde794a40a1d6f26a9657e10412bb: Status 404 returned error can't find the container with id d4f66349559f4a30ec7a570ba461830dcd2cde794a40a1d6f26a9657e10412bb Oct 04 03:42:48 crc kubenswrapper[4726]: W1004 03:42:48.832689 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0444b9ed_94cf_4aab_84a7_f13208c48f64.slice/crio-9119d6d2c92054536f09919ca1f0dd08b3fafb208a8b15313b5b6da6b53963c8 WatchSource:0}: Error finding container 9119d6d2c92054536f09919ca1f0dd08b3fafb208a8b15313b5b6da6b53963c8: Status 404 returned error can't find the container with id 9119d6d2c92054536f09919ca1f0dd08b3fafb208a8b15313b5b6da6b53963c8 Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.835012 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.847002 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.854864 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsq4j\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.859945 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c216c9a6-d84c-42f6-b900-0f0c37c1e107-metrics-tls\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.871561 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.894800 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7jth\" (UniqueName: \"kubernetes.io/projected/42d8eec6-1736-44e1-adfd-995ecbde57a5-kube-api-access-w7jth\") pod \"router-default-5444994796-zx9sv\" (UID: \"42d8eec6-1736-44e1-adfd-995ecbde57a5\") " pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.919995 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:48 crc kubenswrapper[4726]: E1004 03:42:48.920496 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.420479196 +0000 UTC m=+143.595102419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.926173 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht526\" (UniqueName: \"kubernetes.io/projected/9fe3dbf5-b5bb-4445-b876-00398bc99ffe-kube-api-access-ht526\") pod \"ingress-operator-5b745b69d9-qg6rg\" (UID: \"9fe3dbf5-b5bb-4445-b876-00398bc99ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.926300 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lzpf\" (UniqueName: \"kubernetes.io/projected/2950e709-3e91-437c-8f7e-0c85fc073579-kube-api-access-5lzpf\") pod \"catalog-operator-68c6474976-6d56h\" (UID: \"2950e709-3e91-437c-8f7e-0c85fc073579\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.943277 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.974634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bd0f8253-f45f-4567-9c0b-b9ece6299fd0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9zxxm\" (UID: \"bd0f8253-f45f-4567-9c0b-b9ece6299fd0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.990252 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj6c6\" (UniqueName: \"kubernetes.io/projected/503a1e39-10e4-4d8e-946a-410e83081176-kube-api-access-lj6c6\") pod \"package-server-manager-789f6589d5-nw97k\" (UID: \"503a1e39-10e4-4d8e-946a-410e83081176\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:48 crc kubenswrapper[4726]: I1004 03:42:48.991630 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.006974 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.008831 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.012354 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8gtj\" (UniqueName: \"kubernetes.io/projected/72ca42a7-1fe7-4df8-b877-d0d5f94291fe-kube-api-access-x8gtj\") pod \"csi-hostpathplugin-9l5gq\" (UID: \"72ca42a7-1fe7-4df8-b877-d0d5f94291fe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.022016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.022426 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.522411343 +0000 UTC m=+143.697034556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.030380 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg8vk\" (UniqueName: \"kubernetes.io/projected/c216c9a6-d84c-42f6-b900-0f0c37c1e107-kube-api-access-kg8vk\") pod \"dns-default-28vhf\" (UID: \"c216c9a6-d84c-42f6-b900-0f0c37c1e107\") " pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.055848 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ms7\" (UniqueName: \"kubernetes.io/projected/5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d-kube-api-access-p7ms7\") pod \"ingress-canary-nqmms\" (UID: \"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d\") " pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.064491 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.067870 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28ngc\" (UniqueName: \"kubernetes.io/projected/3617abd0-173f-4e18-a1de-8059039e542c-kube-api-access-28ngc\") pod \"machine-config-server-6vzk5\" (UID: \"3617abd0-173f-4e18-a1de-8059039e542c\") " pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.075274 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ln2m"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.080092 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.115059 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.125237 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.125447 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.6254132 +0000 UTC m=+143.800036413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.125690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.125961 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.625949434 +0000 UTC m=+143.800572647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.177559 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.207586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.215842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6vzk5" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.222658 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nqmms" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.223297 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.224821 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.226316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.226510 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.726488823 +0000 UTC m=+143.901112056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.226595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.226862 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.726851473 +0000 UTC m=+143.901474706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.233673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.327712 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.327834 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.827807283 +0000 UTC m=+144.002430496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.327889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.328244 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.828232065 +0000 UTC m=+144.002855278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.343274 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" event={"ID":"0444b9ed-94cf-4aab-84a7-f13208c48f64","Type":"ContainerStarted","Data":"9119d6d2c92054536f09919ca1f0dd08b3fafb208a8b15313b5b6da6b53963c8"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.345511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" event={"ID":"1aad645d-c461-40b5-94e4-b5312f0a3ff7","Type":"ContainerStarted","Data":"ba253441af81aa236bb97927efa22c39a50e2b0a046b255d95d8c1bf94de1821"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.347340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-x6s4v" event={"ID":"8cdb834a-e95f-4532-bd81-207fa47529aa","Type":"ContainerStarted","Data":"d639cbcef8aebbbbe7e1efe3a6a3f117e4f8354ba4f870f205b42dcbba66d3f3"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.348364 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" event={"ID":"c2ca287e-027c-44b2-8068-b84437de7e71","Type":"ContainerStarted","Data":"d4f66349559f4a30ec7a570ba461830dcd2cde794a40a1d6f26a9657e10412bb"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.349653 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" event={"ID":"7386c233-34f0-4eef-87d0-ac2b7b145897","Type":"ContainerStarted","Data":"84f9197a27caf4d6eca10dbc59ea473cc2be23e6325893de68552ed87fa46bd7"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.350452 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nw2zq" event={"ID":"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d","Type":"ContainerStarted","Data":"84448593f4575518589387ff7eae34e9d5287bc0fdf0f00d2896a642f4f6a44c"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.352425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" event={"ID":"4c8fbfde-a722-4762-b6f4-b714a96f1c9f","Type":"ContainerStarted","Data":"a86a8429650cb36795ebed359f38580bc554c57a14187b41510f27bdbc819dfa"} Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.352586 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pbtx9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.352637 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.365230 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wp2cj"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.368931 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xxrt5"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.429823 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.430057 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:49.930025548 +0000 UTC m=+144.104648811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.533201 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.533568 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.033548789 +0000 UTC m=+144.208172012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: W1004 03:42:49.566623 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1da19f66_1ae7_4e73_8f6c_84c471711441.slice/crio-719e8f22771402be8fbe2e284e05773f7142a70ec03fa42eb334551b1540958b WatchSource:0}: Error finding container 719e8f22771402be8fbe2e284e05773f7142a70ec03fa42eb334551b1540958b: Status 404 returned error can't find the container with id 719e8f22771402be8fbe2e284e05773f7142a70ec03fa42eb334551b1540958b Oct 04 03:42:49 crc kubenswrapper[4726]: W1004 03:42:49.612695 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod819db0bd_0b6e_407f_bdde_3aaa513d66a2.slice/crio-e3dbee80848e426665e07d27d06be77b882e6e732f48b5c0134fb425bef3aa69 WatchSource:0}: Error finding container e3dbee80848e426665e07d27d06be77b882e6e732f48b5c0134fb425bef3aa69: Status 404 returned error can't find the container with id e3dbee80848e426665e07d27d06be77b882e6e732f48b5c0134fb425bef3aa69 Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.634047 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.634378 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.134350235 +0000 UTC m=+144.308973468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.634588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.634868 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.134858669 +0000 UTC m=+144.309481882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.730312 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.735010 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.735381 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.235367326 +0000 UTC m=+144.409990539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.836091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.836732 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.336713327 +0000 UTC m=+144.511336540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.862809 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2277j"] Oct 04 03:42:49 crc kubenswrapper[4726]: I1004 03:42:49.936765 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:49 crc kubenswrapper[4726]: E1004 03:42:49.937128 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.437115442 +0000 UTC m=+144.611738655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.042219 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.042539 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.542528515 +0000 UTC m=+144.717151728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.116960 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-h77cw"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.138696 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.145596 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.145883 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.645869871 +0000 UTC m=+144.820493084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.226304 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3617abd0_173f_4e18_a1de_8059039e542c.slice/crio-1f8268ab8052742529de3b49e289c322e582763c0f81e2a809c0fe41a9e1bff1 WatchSource:0}: Error finding container 1f8268ab8052742529de3b49e289c322e582763c0f81e2a809c0fe41a9e1bff1: Status 404 returned error can't find the container with id 1f8268ab8052742529de3b49e289c322e582763c0f81e2a809c0fe41a9e1bff1 Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.248464 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.248759 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.748748354 +0000 UTC m=+144.923371567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.264905 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod654dedf3_d44a_4b22_bce8_6623f832a8f4.slice/crio-dff5d201a467c48402c00af880d85a706583a2f33705914f0730634deab5a51f WatchSource:0}: Error finding container dff5d201a467c48402c00af880d85a706583a2f33705914f0730634deab5a51f: Status 404 returned error can't find the container with id dff5d201a467c48402c00af880d85a706583a2f33705914f0730634deab5a51f Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.352513 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.353009 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.852992575 +0000 UTC m=+145.027615788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.361564 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.409575 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.415473 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mhtxp"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.416204 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" event={"ID":"02db15c4-f787-4ed7-be15-bf9f7e59b342","Type":"ContainerStarted","Data":"3c0c96ac6b1f2f5d91b8d326b7778b2f9e5c6bfb176f22ee23988a1e31faa82d"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.419325 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8mmdc"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.419353 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.425033 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.426329 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nw2zq" event={"ID":"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d","Type":"ContainerStarted","Data":"158844b1e06cfd5e9a24e38f604cac481dd2a469d7017774464add6666343d4e"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.426580 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.427791 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" event={"ID":"1828b5f3-26e7-4b77-aefb-e25c713d7a37","Type":"ContainerStarted","Data":"1dadfcfab0af81c5ce326e7017752c984361ccf44ad3caf68555ecfaa3309f55"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.427851 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.427887 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.429560 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" podStartSLOduration=122.42954844 podStartE2EDuration="2m2.42954844s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:50.402047211 +0000 UTC m=+144.576670424" watchObservedRunningTime="2025-10-04 03:42:50.42954844 +0000 UTC m=+144.604171653" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.430963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" event={"ID":"f284fb82-6924-4615-b695-af1d7de0ec31","Type":"ContainerStarted","Data":"1ba8a0f05f286ca75a7ec2f26b9bd9ac89c5c231c29284c6cf32d2c286f0b66d"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.432561 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" event={"ID":"819db0bd-0b6e-407f-bdde-3aaa513d66a2","Type":"ContainerStarted","Data":"e3dbee80848e426665e07d27d06be77b882e6e732f48b5c0134fb425bef3aa69"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.448360 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" event={"ID":"c2ca287e-027c-44b2-8068-b84437de7e71","Type":"ContainerStarted","Data":"5e78fa72a941428ddbc806f3eebc4904263cb430ef50f4cb7a4cbefeb19b9562"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.449049 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.450133 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ttws2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.450180 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.453243 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" event={"ID":"d0883b34-7264-4f42-b041-c8bb3b742a3d","Type":"ContainerStarted","Data":"32a21ad333b35aa1011a90fb6a212e974814b5ee7431f357d2ac727b03a0669c"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.454019 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.456327 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:50.954425488 +0000 UTC m=+145.129048701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.468469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-x6s4v" event={"ID":"8cdb834a-e95f-4532-bd81-207fa47529aa","Type":"ContainerStarted","Data":"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.470134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" event={"ID":"b5c64dda-8b0b-404e-a875-037d3d7becb3","Type":"ContainerStarted","Data":"895bebdc7a717c53889b3b753d80f08d6f8277d1fa0f15c07d8f377df6ec4b0d"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.490889 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6vzk5" event={"ID":"3617abd0-173f-4e18-a1de-8059039e542c","Type":"ContainerStarted","Data":"1f8268ab8052742529de3b49e289c322e582763c0f81e2a809c0fe41a9e1bff1"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.500431 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" event={"ID":"4c8fbfde-a722-4762-b6f4-b714a96f1c9f","Type":"ContainerStarted","Data":"9530dcbec750bc4e4c428589d3f9d8bc6e54001cc126f36de023bec34ce5796e"} Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.509252 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d91c093_6638_48e4_aa66_6b096cb68cf9.slice/crio-a899f08a6f7bf125b3574a9b2d69fd34a03161015847ee39ac7ee8d5a7b488f7 WatchSource:0}: Error finding container a899f08a6f7bf125b3574a9b2d69fd34a03161015847ee39ac7ee8d5a7b488f7: Status 404 returned error can't find the container with id a899f08a6f7bf125b3574a9b2d69fd34a03161015847ee39ac7ee8d5a7b488f7 Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.510297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" event={"ID":"fa73eb1e-f78b-41f1-9be1-aeae979d66f9","Type":"ContainerStarted","Data":"b8eeddce5dca95c4c372f0f3c6be78f03a66d1cc8e868b8a8823d5c6d13eed2e"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.517336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" event={"ID":"0444b9ed-94cf-4aab-84a7-f13208c48f64","Type":"ContainerStarted","Data":"8b7a10827c460a1853efaa257e44f9e0b86f49583fa9a209aed829ec32665841"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.522404 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.522989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" event={"ID":"1d249e87-e695-4058-8b58-a068fcd85857","Type":"ContainerStarted","Data":"10cfd0366c7d1692c742e3bfbbf050de003a8daa1ed47032313dfa8bdc81e4f1"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.523570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" event={"ID":"0024909c-8d97-45bc-bcd8-a2dba7a101cb","Type":"ContainerStarted","Data":"85d60648877cfa25c0ce63861ec8095b7d4a227b0cef20b4438a89e0c2b43db2"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.533358 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" event={"ID":"1da19f66-1ae7-4e73-8f6c-84c471711441","Type":"ContainerStarted","Data":"719e8f22771402be8fbe2e284e05773f7142a70ec03fa42eb334551b1540958b"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.539054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zx9sv" event={"ID":"42d8eec6-1736-44e1-adfd-995ecbde57a5","Type":"ContainerStarted","Data":"5d2462133f87c29d12c8568e395102abf89491fecf642d337bf985880a704d0c"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.542597 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" event={"ID":"654dedf3-d44a-4b22-bce8-6623f832a8f4","Type":"ContainerStarted","Data":"dff5d201a467c48402c00af880d85a706583a2f33705914f0730634deab5a51f"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.546460 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" event={"ID":"7386c233-34f0-4eef-87d0-ac2b7b145897","Type":"ContainerStarted","Data":"fe0de184d767c316f82280577a51c61ee57b7571ea7611485d89160e6a8c7b8f"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.548944 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" event={"ID":"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9","Type":"ContainerStarted","Data":"d8f0f520f8ab95e330b161bd55e677738a608d35d6f5d5d5711ed166e01e1459"} Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.555741 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.555918 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.055876371 +0000 UTC m=+145.230516345 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.556415 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.556889 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.056829008 +0000 UTC m=+145.231452221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.657256 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.657707 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.658459 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.158441206 +0000 UTC m=+145.333064429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.661878 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.667878 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.670074 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-28vhf"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.676058 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.677255 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.719132 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l5gq"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.723338 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg"] Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.737126 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6cc271a_fe0c_40ca_a28d_68568dcdf2d2.slice/crio-b9e0eee9a823ee55a16e205aba87ae059038fd85f635d825e6948f893abe2179 WatchSource:0}: Error finding container b9e0eee9a823ee55a16e205aba87ae059038fd85f635d825e6948f893abe2179: Status 404 returned error can't find the container with id b9e0eee9a823ee55a16e205aba87ae059038fd85f635d825e6948f893abe2179 Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.741493 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ddfe3ae_32c0_4636_8a34_87b418aa209a.slice/crio-b6474cecbc9cee362b4242b58fd073c43e713ba7ab361b93f51f9ce2a58d8115 WatchSource:0}: Error finding container b6474cecbc9cee362b4242b58fd073c43e713ba7ab361b93f51f9ce2a58d8115: Status 404 returned error can't find the container with id b6474cecbc9cee362b4242b58fd073c43e713ba7ab361b93f51f9ce2a58d8115 Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.746026 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.752264 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7"] Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.753088 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72ca42a7_1fe7_4df8_b877_d0d5f94291fe.slice/crio-c99fa5d4c08e7acd251df50f2fe165b3d82a291dc2af016c83eaf26f31db5ddc WatchSource:0}: Error finding container c99fa5d4c08e7acd251df50f2fe165b3d82a291dc2af016c83eaf26f31db5ddc: Status 404 returned error can't find the container with id c99fa5d4c08e7acd251df50f2fe165b3d82a291dc2af016c83eaf26f31db5ddc Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.765565 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-nw2zq" podStartSLOduration=122.765546256 podStartE2EDuration="2m2.765546256s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:50.765033502 +0000 UTC m=+144.939656725" watchObservedRunningTime="2025-10-04 03:42:50.765546256 +0000 UTC m=+144.940169469" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.773898 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.774902 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.274887374 +0000 UTC m=+145.449510587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.818890 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-t6rc4" podStartSLOduration=122.81887542 podStartE2EDuration="2m2.81887542s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:50.817820961 +0000 UTC m=+144.992444174" watchObservedRunningTime="2025-10-04 03:42:50.81887542 +0000 UTC m=+144.993498633" Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.849494 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nqmms"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.851910 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm"] Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.879117 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.879371 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.379354011 +0000 UTC m=+145.553977214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.879493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.879799 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.379792363 +0000 UTC m=+145.554415576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.925936 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd0f8253_f45f_4567_9c0b_b9ece6299fd0.slice/crio-f89eae63619cb32be8c073aa5081eb1a971052db4176885fd4cef2e4d2c7d09a WatchSource:0}: Error finding container f89eae63619cb32be8c073aa5081eb1a971052db4176885fd4cef2e4d2c7d09a: Status 404 returned error can't find the container with id f89eae63619cb32be8c073aa5081eb1a971052db4176885fd4cef2e4d2c7d09a Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.926900 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podStartSLOduration=122.926881515 podStartE2EDuration="2m2.926881515s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:50.921031833 +0000 UTC m=+145.095655046" watchObservedRunningTime="2025-10-04 03:42:50.926881515 +0000 UTC m=+145.101504728" Oct 04 03:42:50 crc kubenswrapper[4726]: W1004 03:42:50.933424 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b79dafb_a2b8_4f4f_9833_15f4b0a4b39d.slice/crio-5012d436c4b06ec1d8d693103cf4912d5ede57fc5b9a4b13c97af049ccfd655f WatchSource:0}: Error finding container 5012d436c4b06ec1d8d693103cf4912d5ede57fc5b9a4b13c97af049ccfd655f: Status 404 returned error can't find the container with id 5012d436c4b06ec1d8d693103cf4912d5ede57fc5b9a4b13c97af049ccfd655f Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.980832 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.981053 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.481021961 +0000 UTC m=+145.655645194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:50 crc kubenswrapper[4726]: I1004 03:42:50.981370 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:50 crc kubenswrapper[4726]: E1004 03:42:50.981720 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.4817076 +0000 UTC m=+145.656330833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.082406 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.082815 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.582790143 +0000 UTC m=+145.757413356 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.148804 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.183971 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.184320 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.684308019 +0000 UTC m=+145.858931232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.285267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.285438 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.785413973 +0000 UTC m=+145.960037186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.285719 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.286352 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.786335809 +0000 UTC m=+145.960959022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.387985 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.388169 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.888087281 +0000 UTC m=+146.062710494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.388734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.389097 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.889086988 +0000 UTC m=+146.063710201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.490289 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.490457 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.990436849 +0000 UTC m=+146.165060062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.490727 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.490976 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:51.990965384 +0000 UTC m=+146.165588587 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.570569 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" event={"ID":"1622c3b3-4111-4b64-9c2a-4b1a018ff536","Type":"ContainerStarted","Data":"d425231054fc2037b132d789bc593d94d40c73ea71d9f3331d8e066eca710724"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.582871 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" event={"ID":"607637ee-8700-4690-9d42-0a7618d85c73","Type":"ContainerStarted","Data":"502971da5b12ffaaf6c1bdb0ab82925894339a0925a519e24e9c37669e4b05dc"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.588277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" event={"ID":"3a92e571-ee54-4832-a6f5-c9409da8b83d","Type":"ContainerStarted","Data":"a02aeadabe9866018dbd3b27cea2e7cb0075a53f6ad5499f659ec7af0f0cb94a"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.591938 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.592288 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.092274074 +0000 UTC m=+146.266897287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.618650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" event={"ID":"7b069e2c-070d-4051-85b6-3af46fb2e22c","Type":"ContainerStarted","Data":"822705fcf39dd9daee1f06207342d82786a5d63cab73a9cd636b9098e3531ab2"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.623555 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" event={"ID":"72ca42a7-1fe7-4df8-b877-d0d5f94291fe","Type":"ContainerStarted","Data":"c99fa5d4c08e7acd251df50f2fe165b3d82a291dc2af016c83eaf26f31db5ddc"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.626199 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-28vhf" event={"ID":"c216c9a6-d84c-42f6-b900-0f0c37c1e107","Type":"ContainerStarted","Data":"9a2a08eb8a546251ae9171016b9a2f63e21866ce6c37865da7c780a9b12117cd"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.640962 4726 generic.go:334] "Generic (PLEG): container finished" podID="0444b9ed-94cf-4aab-84a7-f13208c48f64" containerID="8b7a10827c460a1853efaa257e44f9e0b86f49583fa9a209aed829ec32665841" exitCode=0 Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.641035 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" event={"ID":"0444b9ed-94cf-4aab-84a7-f13208c48f64","Type":"ContainerDied","Data":"8b7a10827c460a1853efaa257e44f9e0b86f49583fa9a209aed829ec32665841"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.643270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" event={"ID":"9d91c093-6638-48e4-aa66-6b096cb68cf9","Type":"ContainerStarted","Data":"a899f08a6f7bf125b3574a9b2d69fd34a03161015847ee39ac7ee8d5a7b488f7"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.648154 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" event={"ID":"1828b5f3-26e7-4b77-aefb-e25c713d7a37","Type":"ContainerStarted","Data":"1e18271e40b8f5f085837cf721725c58f5a27ccaaf8ff480ef9294df32809ab3"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.649157 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" event={"ID":"4c620617-c3fe-4350-aa40-869d930d9362","Type":"ContainerStarted","Data":"45e1aeaacec17796026305dbc097691893178059342d76b51926bf084284da0c"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.649924 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" event={"ID":"2950e709-3e91-437c-8f7e-0c85fc073579","Type":"ContainerStarted","Data":"4b29df8211a1abe298d3ee342d1070de47f4be629f84ea747214be69552c6c72"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.655051 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" event={"ID":"7179a374-528c-4a2f-a648-998a7a5da73e","Type":"ContainerStarted","Data":"44eed5c708e332e9e2770a84e1dca6ef22689e17fcd28d2ea8a7856749016187"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.657622 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" event={"ID":"9fe3dbf5-b5bb-4445-b876-00398bc99ffe","Type":"ContainerStarted","Data":"76e0df4163e450249e48c8cb96d91e21bfb4b7cc20feec46af3b812b20c432f6"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.658649 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" event={"ID":"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9","Type":"ContainerStarted","Data":"b7a153cdee7ac3a0d6ad5276e4667d59aefd27ff2374414e533b5468a6dd0fbe"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.661232 4726 generic.go:334] "Generic (PLEG): container finished" podID="1aad645d-c461-40b5-94e4-b5312f0a3ff7" containerID="1ccca65484ccd5083be7a39b89606116405d9cfdb266e85829991157cea23d68" exitCode=0 Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.661302 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" event={"ID":"1aad645d-c461-40b5-94e4-b5312f0a3ff7","Type":"ContainerDied","Data":"1ccca65484ccd5083be7a39b89606116405d9cfdb266e85829991157cea23d68"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.663072 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" event={"ID":"986c01ee-6bf6-4654-a8b4-96ee71ace277","Type":"ContainerStarted","Data":"2c18c37a38b8b85b05ca628d53ba02665f39a0c6f14943ed2ccd5404f26c747e"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.663922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" event={"ID":"7ddfe3ae-32c0-4636-8a34-87b418aa209a","Type":"ContainerStarted","Data":"b6474cecbc9cee362b4242b58fd073c43e713ba7ab361b93f51f9ce2a58d8115"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.665394 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" event={"ID":"01bcb480-958d-4346-b185-44d717844903","Type":"ContainerStarted","Data":"a49379467618c0c855f6e72098557d8f962ad1348e8c57c248ccf2f835965470"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.668138 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" event={"ID":"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00","Type":"ContainerStarted","Data":"2e53121cd8686eb0a7ccff9c974c08f3ea9ea6dcf3b1c19a3214f279ea600626"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.670019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nqmms" event={"ID":"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d","Type":"ContainerStarted","Data":"5012d436c4b06ec1d8d693103cf4912d5ede57fc5b9a4b13c97af049ccfd655f"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.671447 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" event={"ID":"503a1e39-10e4-4d8e-946a-410e83081176","Type":"ContainerStarted","Data":"d5623527742a7074ee4a3eefffdc52ebb55237dfa8ae8fc16154165f784e6241"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.676360 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" event={"ID":"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2","Type":"ContainerStarted","Data":"b9e0eee9a823ee55a16e205aba87ae059038fd85f635d825e6948f893abe2179"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.693525 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.694341 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.194317644 +0000 UTC m=+146.368940857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.700990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" event={"ID":"bd0f8253-f45f-4567-9c0b-b9ece6299fd0","Type":"ContainerStarted","Data":"f89eae63619cb32be8c073aa5081eb1a971052db4176885fd4cef2e4d2c7d09a"} Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.702349 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.702402 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.702503 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ttws2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.702546 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.732979 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-x6s4v" podStartSLOduration=123.732964782 podStartE2EDuration="2m3.732964782s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:51.732568101 +0000 UTC m=+145.907191314" watchObservedRunningTime="2025-10-04 03:42:51.732964782 +0000 UTC m=+145.907587995" Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.794039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.795045 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.295032207 +0000 UTC m=+146.469655420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.895531 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.895806 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.395796272 +0000 UTC m=+146.570419485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.998246 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.998619 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.498604413 +0000 UTC m=+146.673227626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:51 crc kubenswrapper[4726]: I1004 03:42:51.998722 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:51 crc kubenswrapper[4726]: E1004 03:42:51.998980 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.498973503 +0000 UTC m=+146.673596716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.099200 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.099398 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.599367418 +0000 UTC m=+146.773990631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.099713 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.100261 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.600252682 +0000 UTC m=+146.774875895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.200951 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.201155 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.701097549 +0000 UTC m=+146.875720802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.201382 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.201716 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.701702216 +0000 UTC m=+146.876325509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.302285 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.302482 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.80245203 +0000 UTC m=+146.977075263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.302726 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.303193 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.80315045 +0000 UTC m=+146.977773703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.403404 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.403557 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.903530144 +0000 UTC m=+147.078153357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.403667 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.403980 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:52.903970766 +0000 UTC m=+147.078594029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.505151 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.505352 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.005326057 +0000 UTC m=+147.179949270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.505469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.505763 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.005755939 +0000 UTC m=+147.180379142 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.606684 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.607043 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.107025408 +0000 UTC m=+147.281648621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.607306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.607560 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.107552832 +0000 UTC m=+147.282176045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.707776 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.708214 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.208192034 +0000 UTC m=+147.382815257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.709140 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" event={"ID":"4c620617-c3fe-4350-aa40-869d930d9362","Type":"ContainerStarted","Data":"3eae699577d77f259b5489fba1fe6714a876d0697da9b38312f71aaa05313219"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.710831 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" event={"ID":"1d249e87-e695-4058-8b58-a068fcd85857","Type":"ContainerStarted","Data":"edfad6528dff4afdcd90b4b127589d14fd1164b5f9fdb0db13e19971f609cd05"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.712547 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" event={"ID":"819db0bd-0b6e-407f-bdde-3aaa513d66a2","Type":"ContainerStarted","Data":"2ecebefe504e3c4f71de07c9c0985036d365dc98628ffb92ba1d1a147861bdae"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.713877 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.716032 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-wp2cj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.716085 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" podUID="819db0bd-0b6e-407f-bdde-3aaa513d66a2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.717933 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" event={"ID":"986c01ee-6bf6-4654-a8b4-96ee71ace277","Type":"ContainerStarted","Data":"c3fd0960eed7571585e84e88a79e876c02a88e08ed765d2332bc72da2a0f65bf"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.724760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" event={"ID":"7ddfe3ae-32c0-4636-8a34-87b418aa209a","Type":"ContainerStarted","Data":"ea85926dd1fc496bb420ddd6efcb4c5e43a92c851a3331e55effe011c9daa931"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.731177 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" event={"ID":"0024909c-8d97-45bc-bcd8-a2dba7a101cb","Type":"ContainerStarted","Data":"75dc008538539256980d83ae27a4926c24a541c0c1a1520672ae93e691a86d59"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.732509 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-28vhf" event={"ID":"c216c9a6-d84c-42f6-b900-0f0c37c1e107","Type":"ContainerStarted","Data":"da1fd7e2e88f8eeacadc15e214b60c96d013784d9bd09c698632c54623254ec6"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.735130 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" event={"ID":"1622c3b3-4111-4b64-9c2a-4b1a018ff536","Type":"ContainerStarted","Data":"568fb34766fb8058a27f92b3a7d37e9eafacbb338b5195f3f00b2d13fd7118bb"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.738048 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6vzk5" event={"ID":"3617abd0-173f-4e18-a1de-8059039e542c","Type":"ContainerStarted","Data":"d75831f7dd749b7beec9a29b235a6e31b8a8a1d885c46d5d7a9d7ddc5fe6e2a9"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.742564 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" event={"ID":"9fe3dbf5-b5bb-4445-b876-00398bc99ffe","Type":"ContainerStarted","Data":"3eafd8c59d619fedbad26ef73ba613824d3dbe5fd1f876f6039593629aedcc5c"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.745249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" event={"ID":"1da19f66-1ae7-4e73-8f6c-84c471711441","Type":"ContainerStarted","Data":"32e85887976e2c84ae10771767f9c3eefc9a7ccba878d3f7bb7d75edce870079"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.754328 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" event={"ID":"b5c64dda-8b0b-404e-a875-037d3d7becb3","Type":"ContainerStarted","Data":"fdf55af040a3206df8e6ab6bb74acb2b00c50c5b54dfc0643990c23f8141f465"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.758495 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" podStartSLOduration=124.758481523 podStartE2EDuration="2m4.758481523s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.756996242 +0000 UTC m=+146.931619455" watchObservedRunningTime="2025-10-04 03:42:52.758481523 +0000 UTC m=+146.933104736" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.759431 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" event={"ID":"654dedf3-d44a-4b22-bce8-6623f832a8f4","Type":"ContainerStarted","Data":"8b717bcb3eb1b19df2080a8481cf17b5ad9ad095439a42ccc2625887031a32ae"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.760936 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" event={"ID":"3a92e571-ee54-4832-a6f5-c9409da8b83d","Type":"ContainerStarted","Data":"7f3a90077c1be3fb4106440b5111ef8a3186754b36194a2797e3767ca883435e"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.797217 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ctsv5" podStartSLOduration=124.797201813 podStartE2EDuration="2m4.797201813s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.793580643 +0000 UTC m=+146.968203846" watchObservedRunningTime="2025-10-04 03:42:52.797201813 +0000 UTC m=+146.971825026" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.810587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" event={"ID":"9d91c093-6638-48e4-aa66-6b096cb68cf9","Type":"ContainerStarted","Data":"8dde5a46176599b8e91bc9f3f17adadcfb62cd89137de9cfa305b8a829273a09"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.811041 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.811990 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.311975202 +0000 UTC m=+147.486598415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.821482 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gksb9" podStartSLOduration=124.821466384 podStartE2EDuration="2m4.821466384s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.820381964 +0000 UTC m=+146.995005177" watchObservedRunningTime="2025-10-04 03:42:52.821466384 +0000 UTC m=+146.996089597" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.825470 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zx9sv" event={"ID":"42d8eec6-1736-44e1-adfd-995ecbde57a5","Type":"ContainerStarted","Data":"57a37158466df690a1994307ddf3c862e1b6af2b3ab458b70d7150e35c749781"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.832939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" event={"ID":"fa73eb1e-f78b-41f1-9be1-aeae979d66f9","Type":"ContainerStarted","Data":"a6840ea45066b16157118e976b990bb31da7bda942cef6ba942e91825bba4a03"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.836958 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ln2m" podStartSLOduration=124.836941102 podStartE2EDuration="2m4.836941102s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.835566894 +0000 UTC m=+147.010190107" watchObservedRunningTime="2025-10-04 03:42:52.836941102 +0000 UTC m=+147.011564315" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.837882 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" event={"ID":"607637ee-8700-4690-9d42-0a7618d85c73","Type":"ContainerStarted","Data":"89ccfebf842db566c588dead967aac5b9266dfc247bcb1f45848e3996674bb62"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.840996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" event={"ID":"7b069e2c-070d-4051-85b6-3af46fb2e22c","Type":"ContainerStarted","Data":"2f52b0caa9e89424fe79ffd0f9c4f193eb2bef2b25bcdcd17af83635c1bdefcd"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.843502 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" event={"ID":"eb0f3a0d-d4f0-41d0-8c41-f1bf1e891b00","Type":"ContainerStarted","Data":"c836ac9eb5661ef7b4a8bc9f4ecb4891a720ed6eaab28410a0a3561feb19c495"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.848727 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" event={"ID":"2950e709-3e91-437c-8f7e-0c85fc073579","Type":"ContainerStarted","Data":"180c078da65ccbffaa40d7f1633e2568a278d98f0925d99b0d08ef949fd7ee20"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.850866 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" event={"ID":"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2","Type":"ContainerStarted","Data":"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.852233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" event={"ID":"7386c233-34f0-4eef-87d0-ac2b7b145897","Type":"ContainerStarted","Data":"4cab0a15ff4d98ae5edadbeceae1af3ba2453be56f098f867e66aaea42dc5e4a"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.854214 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" event={"ID":"02db15c4-f787-4ed7-be15-bf9f7e59b342","Type":"ContainerStarted","Data":"4e536ee272927c945c1dde7130f6dd177253ea6f39eb23de7a42605a054100ce"} Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.854622 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ttws2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.854650 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.860945 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6vzk5" podStartSLOduration=7.860928235 podStartE2EDuration="7.860928235s" podCreationTimestamp="2025-10-04 03:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.859417273 +0000 UTC m=+147.034040486" watchObservedRunningTime="2025-10-04 03:42:52.860928235 +0000 UTC m=+147.035551438" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.900973 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-h77cw" podStartSLOduration=124.900956471 podStartE2EDuration="2m4.900956471s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.900766186 +0000 UTC m=+147.075389399" watchObservedRunningTime="2025-10-04 03:42:52.900956471 +0000 UTC m=+147.075579684" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.903425 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qstzg" podStartSLOduration=124.903406989 podStartE2EDuration="2m4.903406989s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.877655767 +0000 UTC m=+147.052278980" watchObservedRunningTime="2025-10-04 03:42:52.903406989 +0000 UTC m=+147.078030192" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.914991 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.915181 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.415161623 +0000 UTC m=+147.589784836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.915499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:52 crc kubenswrapper[4726]: E1004 03:42:52.916403 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.416388167 +0000 UTC m=+147.591011380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.932607 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zx9sv" podStartSLOduration=124.932589005 podStartE2EDuration="2m4.932589005s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:52.931927687 +0000 UTC m=+147.106550900" watchObservedRunningTime="2025-10-04 03:42:52.932589005 +0000 UTC m=+147.107212218" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.946239 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.948038 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 04 03:42:52 crc kubenswrapper[4726]: I1004 03:42:52.948095 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.025954 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.027159 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.527096657 +0000 UTC m=+147.701719870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.127435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.127820 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.62780218 +0000 UTC m=+147.802425403 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.229662 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.229993 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.729977524 +0000 UTC m=+147.904600737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.331174 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.331781 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.831767077 +0000 UTC m=+148.006390300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.432722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.433009 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:53.932994994 +0000 UTC m=+148.107618207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.534585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.534847 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.034835389 +0000 UTC m=+148.209458592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.636281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.636461 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.136437137 +0000 UTC m=+148.311060350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.636635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.636973 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.136964691 +0000 UTC m=+148.311587904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.738370 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.738686 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.238672472 +0000 UTC m=+148.413295685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.840238 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.840915 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.340903538 +0000 UTC m=+148.515526741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.860995 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" event={"ID":"4c8fbfde-a722-4762-b6f4-b714a96f1c9f","Type":"ContainerStarted","Data":"8519dd38bbf3b6a49ace7852298452d3355b702b603047733f432100969b4154"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.863417 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" event={"ID":"d0883b34-7264-4f42-b041-c8bb3b742a3d","Type":"ContainerStarted","Data":"c90fca55eba7c9b590390e9269cb4a70cb9b521fdebb2dd50d7a558fe837181d"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.865470 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" event={"ID":"7b069e2c-070d-4051-85b6-3af46fb2e22c","Type":"ContainerStarted","Data":"7b1a73f9ec0f5907e57e176764f19b3d1b0f2ea1362f8039295777a51079e67a"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.867699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" event={"ID":"9fe3dbf5-b5bb-4445-b876-00398bc99ffe","Type":"ContainerStarted","Data":"15b48dc30234263dba4b8bbaeb20b96c78d031f70fd6cf6341300ccd03ab26bb"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.869500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" event={"ID":"7179a374-528c-4a2f-a648-998a7a5da73e","Type":"ContainerStarted","Data":"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.869708 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.870906 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nqmms" event={"ID":"5b79dafb-a2b8-4f4f-9833-15f4b0a4b39d","Type":"ContainerStarted","Data":"6261cb8ab3173725b07fc02cc0373a3afd4b0c74c1ce6a5c5d13435ff831286b"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.871203 4726 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wssnh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.871329 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.872333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-28vhf" event={"ID":"c216c9a6-d84c-42f6-b900-0f0c37c1e107","Type":"ContainerStarted","Data":"e560713681943d89f443c1c19537f6ffa65390de5ac00db333b508f3ad1d47e9"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.872498 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-28vhf" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.873693 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" event={"ID":"01bcb480-958d-4346-b185-44d717844903","Type":"ContainerStarted","Data":"2b8eb549f20239946ef77fd76e164f7692c5d5181679fd886cbb0e9e71e24cb1"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.873856 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.874631 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-w6fz4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.874676 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" podUID="01bcb480-958d-4346-b185-44d717844903" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.876299 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" event={"ID":"c09a4bf0-3a5c-44ff-a5fb-1f81188362e9","Type":"ContainerStarted","Data":"0181692377fcf969c57d1757ede695984062ab9bcc99d537d0a99c3cdff2055a"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.878081 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" event={"ID":"503a1e39-10e4-4d8e-946a-410e83081176","Type":"ContainerStarted","Data":"2b70182a0f4f16fbe015449d241e368acb6587166ca478c55ac43c8fb191ff0c"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.878140 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" event={"ID":"503a1e39-10e4-4d8e-946a-410e83081176","Type":"ContainerStarted","Data":"862e84a908c25abc97042a337ed67eb660975072cd0016e768af195040fe36e5"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.878157 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.879547 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" event={"ID":"bd0f8253-f45f-4567-9c0b-b9ece6299fd0","Type":"ContainerStarted","Data":"1e3e94a1d6adb023982bf66ff0d08bcd728b5b5245c8e287c7498ca123076951"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.881939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" event={"ID":"02db15c4-f787-4ed7-be15-bf9f7e59b342","Type":"ContainerStarted","Data":"edaba7aa9c5fc5523946d132a4b236c4670a5959765e9e1bd820efb0249fe2bc"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.885266 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" event={"ID":"0444b9ed-94cf-4aab-84a7-f13208c48f64","Type":"ContainerStarted","Data":"625b8bde47134acc2a4de25772a5009c6cd1cb76e936ac7970b6295b21f754d5"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.885419 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.887386 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" event={"ID":"1828b5f3-26e7-4b77-aefb-e25c713d7a37","Type":"ContainerStarted","Data":"bedf543f48ff0187d93beda10810834e12689dabe70ad4a3490c24f9ec60c7c3"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.889245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" event={"ID":"3a92e571-ee54-4832-a6f5-c9409da8b83d","Type":"ContainerStarted","Data":"828200a5aa520785a2c78666a6dd688e0c11fa308622c30e5ce84093c0ebaf5d"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.891270 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-s7ggk" podStartSLOduration=125.891256569 podStartE2EDuration="2m5.891256569s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:53.889157201 +0000 UTC m=+148.063780414" watchObservedRunningTime="2025-10-04 03:42:53.891256569 +0000 UTC m=+148.065879782" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.891473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" event={"ID":"1aad645d-c461-40b5-94e4-b5312f0a3ff7","Type":"ContainerStarted","Data":"c6df1d9216fd74b3c63f2e3da2fd79c89b60a738a02d4eae60778ce4ddad3d8e"} Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.892666 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-wp2cj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.892703 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" podUID="819db0bd-0b6e-407f-bdde-3aaa513d66a2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.892714 4726 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-kcv7c container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.892749 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" podUID="607637ee-8700-4690-9d42-0a7618d85c73" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.893412 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.917003 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" podStartSLOduration=125.9169874 podStartE2EDuration="2m5.9169874s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:53.916205469 +0000 UTC m=+148.090828682" watchObservedRunningTime="2025-10-04 03:42:53.9169874 +0000 UTC m=+148.091610613" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.941946 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.942178 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.442149866 +0000 UTC m=+148.616773079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.942654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:53 crc kubenswrapper[4726]: E1004 03:42:53.942977 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.442969168 +0000 UTC m=+148.617592381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.945347 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.945393 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.951905 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7ksbd" podStartSLOduration=125.951887125 podStartE2EDuration="2m5.951887125s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:53.949803687 +0000 UTC m=+148.124426900" watchObservedRunningTime="2025-10-04 03:42:53.951887125 +0000 UTC m=+148.126510338" Oct 04 03:42:53 crc kubenswrapper[4726]: I1004 03:42:53.981215 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" podStartSLOduration=126.981090602 podStartE2EDuration="2m6.981090602s" podCreationTimestamp="2025-10-04 03:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:53.979821477 +0000 UTC m=+148.154444690" watchObservedRunningTime="2025-10-04 03:42:53.981090602 +0000 UTC m=+148.155713815" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.009874 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" podStartSLOduration=126.009858246 podStartE2EDuration="2m6.009858246s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.008402116 +0000 UTC m=+148.183025349" watchObservedRunningTime="2025-10-04 03:42:54.009858246 +0000 UTC m=+148.184481449" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.031678 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" podStartSLOduration=126.031662138 podStartE2EDuration="2m6.031662138s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.028977314 +0000 UTC m=+148.203600527" watchObservedRunningTime="2025-10-04 03:42:54.031662138 +0000 UTC m=+148.206285351" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.044411 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.047085 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.547070014 +0000 UTC m=+148.721693217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.103372 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-qg6rg" podStartSLOduration=126.10335545 podStartE2EDuration="2m6.10335545s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.099920665 +0000 UTC m=+148.274543878" watchObservedRunningTime="2025-10-04 03:42:54.10335545 +0000 UTC m=+148.277978663" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.150569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.150845 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.650833862 +0000 UTC m=+148.825457075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.176152 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9zxxm" podStartSLOduration=126.176135031 podStartE2EDuration="2m6.176135031s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.174502076 +0000 UTC m=+148.349125289" watchObservedRunningTime="2025-10-04 03:42:54.176135031 +0000 UTC m=+148.350758244" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.226048 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" podStartSLOduration=126.22603526 podStartE2EDuration="2m6.22603526s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.224516698 +0000 UTC m=+148.399139911" watchObservedRunningTime="2025-10-04 03:42:54.22603526 +0000 UTC m=+148.400658473" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.252161 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.252581 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.752566083 +0000 UTC m=+148.927189296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.354449 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.354576 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xxrt5" podStartSLOduration=126.354560332 podStartE2EDuration="2m6.354560332s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.296052765 +0000 UTC m=+148.470675978" watchObservedRunningTime="2025-10-04 03:42:54.354560332 +0000 UTC m=+148.529183545" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.354964 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.854952013 +0000 UTC m=+149.029575226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.354964 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-28vhf" podStartSLOduration=9.354953033 podStartE2EDuration="9.354953033s" podCreationTimestamp="2025-10-04 03:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.342592621 +0000 UTC m=+148.517215834" watchObservedRunningTime="2025-10-04 03:42:54.354953033 +0000 UTC m=+148.529576246" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.372257 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2277j" podStartSLOduration=126.372238271 podStartE2EDuration="2m6.372238271s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.3700401 +0000 UTC m=+148.544663313" watchObservedRunningTime="2025-10-04 03:42:54.372238271 +0000 UTC m=+148.546861484" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.454085 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nqmms" podStartSLOduration=8.454065182 podStartE2EDuration="8.454065182s" podCreationTimestamp="2025-10-04 03:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.41636771 +0000 UTC m=+148.590990923" watchObservedRunningTime="2025-10-04 03:42:54.454065182 +0000 UTC m=+148.628688395" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.455944 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.456322 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:54.956310024 +0000 UTC m=+149.130933237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.456921 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-s6fzl" podStartSLOduration=126.456912451 podStartE2EDuration="2m6.456912451s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.453315641 +0000 UTC m=+148.627938864" watchObservedRunningTime="2025-10-04 03:42:54.456912451 +0000 UTC m=+148.631535664" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.552551 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" podStartSLOduration=126.552533563 podStartE2EDuration="2m6.552533563s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.55131988 +0000 UTC m=+148.725943093" watchObservedRunningTime="2025-10-04 03:42:54.552533563 +0000 UTC m=+148.727156776" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.557062 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.557403 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.057391438 +0000 UTC m=+149.232014651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.597576 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-stsdb" podStartSLOduration=126.597558298 podStartE2EDuration="2m6.597558298s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.570457389 +0000 UTC m=+148.745080602" watchObservedRunningTime="2025-10-04 03:42:54.597558298 +0000 UTC m=+148.772181501" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.623338 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tspxk" podStartSLOduration=126.6233222 podStartE2EDuration="2m6.6233222s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.621746026 +0000 UTC m=+148.796369239" watchObservedRunningTime="2025-10-04 03:42:54.6233222 +0000 UTC m=+148.797945403" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.623957 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-mhtxp" podStartSLOduration=126.623947987 podStartE2EDuration="2m6.623947987s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.598953396 +0000 UTC m=+148.773576609" watchObservedRunningTime="2025-10-04 03:42:54.623947987 +0000 UTC m=+148.798571210" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.659960 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.660162 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.160133157 +0000 UTC m=+149.334756370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.660304 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.660601 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.16059117 +0000 UTC m=+149.335214383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.688754 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xffj7" podStartSLOduration=126.688738968 podStartE2EDuration="2m6.688738968s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.659331705 +0000 UTC m=+148.833954918" watchObservedRunningTime="2025-10-04 03:42:54.688738968 +0000 UTC m=+148.863362181" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.716534 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-58k7b" podStartSLOduration=126.716518705 podStartE2EDuration="2m6.716518705s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.689600071 +0000 UTC m=+148.864223284" watchObservedRunningTime="2025-10-04 03:42:54.716518705 +0000 UTC m=+148.891141918" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.738087 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" podStartSLOduration=126.738070791 podStartE2EDuration="2m6.738070791s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.719248361 +0000 UTC m=+148.893871564" watchObservedRunningTime="2025-10-04 03:42:54.738070791 +0000 UTC m=+148.912694004" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.760947 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.761079 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.261058886 +0000 UTC m=+149.435682099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.761142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.761485 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.261474248 +0000 UTC m=+149.436097461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.762713 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8mmdc" podStartSLOduration=126.762701032 podStartE2EDuration="2m6.762701032s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.760761158 +0000 UTC m=+148.935384371" watchObservedRunningTime="2025-10-04 03:42:54.762701032 +0000 UTC m=+148.937324245" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.763628 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" podStartSLOduration=126.763623897 podStartE2EDuration="2m6.763623897s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.738784961 +0000 UTC m=+148.913408174" watchObservedRunningTime="2025-10-04 03:42:54.763623897 +0000 UTC m=+148.938247110" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.779928 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-xrfvj" podStartSLOduration=126.779914687 podStartE2EDuration="2m6.779914687s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.777936483 +0000 UTC m=+148.952559696" watchObservedRunningTime="2025-10-04 03:42:54.779914687 +0000 UTC m=+148.954537900" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.840982 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ffl97" podStartSLOduration=126.840964375 podStartE2EDuration="2m6.840964375s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.802845571 +0000 UTC m=+148.977468774" watchObservedRunningTime="2025-10-04 03:42:54.840964375 +0000 UTC m=+149.015587588" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.861705 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.862054 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.362039747 +0000 UTC m=+149.536662960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.868125 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" podStartSLOduration=126.868094404 podStartE2EDuration="2m6.868094404s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.841845569 +0000 UTC m=+149.016468782" watchObservedRunningTime="2025-10-04 03:42:54.868094404 +0000 UTC m=+149.042717617" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.896330 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" event={"ID":"72ca42a7-1fe7-4df8-b877-d0d5f94291fe","Type":"ContainerStarted","Data":"1565710ccdf804e43d2825724d9ea1b6cd937d6a7b768958de4dff24da006d5d"} Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.898875 4726 patch_prober.go:28] interesting pod/console-operator-58897d9998-wp2cj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.898910 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" podUID="819db0bd-0b6e-407f-bdde-3aaa513d66a2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.911845 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-kcv7c" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.941293 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" podStartSLOduration=126.941279377 podStartE2EDuration="2m6.941279377s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:54.870190572 +0000 UTC m=+149.044813785" watchObservedRunningTime="2025-10-04 03:42:54.941279377 +0000 UTC m=+149.115902590" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.949876 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:54 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:54 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:54 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.949923 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:54 crc kubenswrapper[4726]: I1004 03:42:54.963243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:54 crc kubenswrapper[4726]: E1004 03:42:54.974894 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.474875185 +0000 UTC m=+149.649498398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.064543 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.064774 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.564740759 +0000 UTC m=+149.739363972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.064890 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.065289 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.565273674 +0000 UTC m=+149.739896887 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.166294 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.166491 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.66646204 +0000 UTC m=+149.841085253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.166880 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.167232 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.667221161 +0000 UTC m=+149.841844444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.268285 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.268480 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.768449669 +0000 UTC m=+149.943072882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.268832 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.269145 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.769133878 +0000 UTC m=+149.943757091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.370298 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.370494 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.870467538 +0000 UTC m=+150.045090751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.370567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.370851 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.870839008 +0000 UTC m=+150.045462221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.471645 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.471816 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.971788378 +0000 UTC m=+150.146411591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.471873 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.471906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.471946 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.471981 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.472013 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.472284 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:55.972272392 +0000 UTC m=+150.146895605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.474385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.477797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.479218 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.489226 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.525665 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.541006 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.569314 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.572732 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.573054 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.073038146 +0000 UTC m=+150.247661359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.674382 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.674689 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.174677745 +0000 UTC m=+150.349300958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.776632 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.776954 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.276940952 +0000 UTC m=+150.451564165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.879366 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.879645 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.37963328 +0000 UTC m=+150.554256493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.899245 4726 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-w6fz4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.899301 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" podUID="01bcb480-958d-4346-b185-44d717844903" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.900239 4726 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wssnh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.900296 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.919012 4726 generic.go:334] "Generic (PLEG): container finished" podID="1d249e87-e695-4058-8b58-a068fcd85857" containerID="edfad6528dff4afdcd90b4b127589d14fd1164b5f9fdb0db13e19971f609cd05" exitCode=0 Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.919078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" event={"ID":"1d249e87-e695-4058-8b58-a068fcd85857","Type":"ContainerDied","Data":"edfad6528dff4afdcd90b4b127589d14fd1164b5f9fdb0db13e19971f609cd05"} Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.980468 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:55 crc kubenswrapper[4726]: E1004 03:42:55.980747 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.480722343 +0000 UTC m=+150.655345556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.998610 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:55 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:55 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:55 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:55 crc kubenswrapper[4726]: I1004 03:42:55.998938 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.087362 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.087771 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.587753621 +0000 UTC m=+150.762376834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.188163 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.188442 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.688417753 +0000 UTC m=+150.863040966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.188745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.189055 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.689048351 +0000 UTC m=+150.863671564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.289749 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.290338 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.79032398 +0000 UTC m=+150.964947193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.392178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.392559 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.892544105 +0000 UTC m=+151.067167318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.403525 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.404580 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.409235 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.431223 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.493512 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.493733 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxtvd\" (UniqueName: \"kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.493782 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.493813 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.493935 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:56.993918976 +0000 UTC m=+151.168542189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.558391 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.562494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.581576 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.586893 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.587784 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.589095 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.597347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.597390 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.597441 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.597460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxtvd\" (UniqueName: \"kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.598257 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.598461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.598661 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.098651451 +0000 UTC m=+151.273274664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.607050 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.632762 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxtvd\" (UniqueName: \"kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd\") pod \"certified-operators-khfd7\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.699039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.699213 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.699247 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.699316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktbbt\" (UniqueName: \"kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.702825 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.202805349 +0000 UTC m=+151.377428562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.711476 4726 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2s6pg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]log ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]etcd ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/max-in-flight-filter ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 04 03:42:56 crc kubenswrapper[4726]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 04 03:42:56 crc kubenswrapper[4726]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectcache ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-startinformers ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 04 03:42:56 crc kubenswrapper[4726]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 03:42:56 crc kubenswrapper[4726]: livez check failed Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.711707 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" podUID="d0883b34-7264-4f42-b041-c8bb3b742a3d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.720071 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.784829 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.785924 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.800628 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.800678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktbbt\" (UniqueName: \"kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.800728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.800758 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.801133 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.801353 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.301342882 +0000 UTC m=+151.475966105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.801541 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.801926 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.827935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktbbt\" (UniqueName: \"kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt\") pod \"community-operators-swwmz\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.902286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.902449 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.402423296 +0000 UTC m=+151.577046509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.902608 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.902675 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwr4g\" (UniqueName: \"kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.902723 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.902824 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:56 crc kubenswrapper[4726]: E1004 03:42:56.903091 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.403081504 +0000 UTC m=+151.577704717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.920716 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.941488 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5e271be2951044ea42bc739c451d5f1d2db1c40ab416f22306fea1fa97f11440"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.941540 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d134998e7f60e362736f83ee00a3087cd5ef6778ef3d833725b7a8ae4c5304ea"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.948305 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:56 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:56 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:56 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.948342 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.948822 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.954587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"548353cf6a429ec7d4dc1e7dfefa4ec16bda5ca6bffd34834afe5d19d891d810"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.954623 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7689513d7dc2700bd80898192eadab2986a2a5939095f8216e9e4007e429901b"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.957556 4726 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.957802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" event={"ID":"72ca42a7-1fe7-4df8-b877-d0d5f94291fe","Type":"ContainerStarted","Data":"138ac5750113d8611d664235aaafced2b2510d916e087f26cb2c2c04390e5f56"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.957828 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" event={"ID":"72ca42a7-1fe7-4df8-b877-d0d5f94291fe","Type":"ContainerStarted","Data":"2e6507a480b2386b7df30d11a7fb0c45f3f3e31af41995425e8403c62be00f22"} Oct 04 03:42:56 crc kubenswrapper[4726]: W1004 03:42:56.958463 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8435186_bbfc_4907_bf0a_2a6b88613f07.slice/crio-d457e3d7c62d68d3d2b5a65761d490be51139beea124181183ff7a5dc944d26a WatchSource:0}: Error finding container d457e3d7c62d68d3d2b5a65761d490be51139beea124181183ff7a5dc944d26a: Status 404 returned error can't find the container with id d457e3d7c62d68d3d2b5a65761d490be51139beea124181183ff7a5dc944d26a Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.967423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4fd8b61bd494194a48c6fc91dbba0e020f4170e8c719feacd21ec4223110b4d9"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.967451 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"67b8a518a9345f2e1772f4d5d946437948345b393b2dd8347782941d7396138c"} Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.967727 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.984664 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:42:56 crc kubenswrapper[4726]: I1004 03:42:56.985598 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.004809 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.004986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.005082 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.005113 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwr4g\" (UniqueName: \"kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.005714 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.005719 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.50569344 +0000 UTC m=+151.680316663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.005931 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.010141 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.074827 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwr4g\" (UniqueName: \"kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g\") pod \"certified-operators-qzrv5\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.112831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.115014 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.115425 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvthp\" (UniqueName: \"kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.115470 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.115503 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.115757 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.615741981 +0000 UTC m=+151.790365194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.127727 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cptzs" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.220753 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.220919 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.720889297 +0000 UTC m=+151.895512510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.221005 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.221091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.221255 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.221367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvthp\" (UniqueName: \"kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.222868 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.223348 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.223588 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.723576751 +0000 UTC m=+151.898199964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.248527 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvthp\" (UniqueName: \"kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp\") pod \"community-operators-vm256\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.319063 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.324761 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.325138 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.825116587 +0000 UTC m=+151.999739800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.330410 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.373223 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:42:57 crc kubenswrapper[4726]: W1004 03:42:57.380120 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9ba01a7_5230_4c26_89e4_19cf823c3ded.slice/crio-f497f2f5cbd9de7024a2ab0118f4544b32bde1ddf6fff1d6534343d6e130a6c6 WatchSource:0}: Error finding container f497f2f5cbd9de7024a2ab0118f4544b32bde1ddf6fff1d6534343d6e130a6c6: Status 404 returned error can't find the container with id f497f2f5cbd9de7024a2ab0118f4544b32bde1ddf6fff1d6534343d6e130a6c6 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.394203 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.413491 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.416338 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d249e87-e695-4058-8b58-a068fcd85857" containerName="collect-profiles" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.416374 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d249e87-e695-4058-8b58-a068fcd85857" containerName="collect-profiles" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.416487 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d249e87-e695-4058-8b58-a068fcd85857" containerName="collect-profiles" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.416980 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.418768 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.421641 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.422491 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.425712 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume\") pod \"1d249e87-e695-4058-8b58-a068fcd85857\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.425773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd2zj\" (UniqueName: \"kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj\") pod \"1d249e87-e695-4058-8b58-a068fcd85857\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.425825 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume\") pod \"1d249e87-e695-4058-8b58-a068fcd85857\" (UID: \"1d249e87-e695-4058-8b58-a068fcd85857\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.426149 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.426455 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:57.926438078 +0000 UTC m=+152.101061291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.427085 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume" (OuterVolumeSpecName: "config-volume") pod "1d249e87-e695-4058-8b58-a068fcd85857" (UID: "1d249e87-e695-4058-8b58-a068fcd85857"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.430776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj" (OuterVolumeSpecName: "kube-api-access-qd2zj") pod "1d249e87-e695-4058-8b58-a068fcd85857" (UID: "1d249e87-e695-4058-8b58-a068fcd85857"). InnerVolumeSpecName "kube-api-access-qd2zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.431022 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1d249e87-e695-4058-8b58-a068fcd85857" (UID: "1d249e87-e695-4058-8b58-a068fcd85857"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.526763 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.527834 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:58.027809549 +0000 UTC m=+152.202432762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528464 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528542 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528565 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528671 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1d249e87-e695-4058-8b58-a068fcd85857-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528685 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd2zj\" (UniqueName: \"kubernetes.io/projected/1d249e87-e695-4058-8b58-a068fcd85857-kube-api-access-qd2zj\") on node \"crc\" DevicePath \"\"" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.528698 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1d249e87-e695-4058-8b58-a068fcd85857-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.528803 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 03:42:58.028787496 +0000 UTC m=+152.203410709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-gq29w" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.558562 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod747e7096_4e73_42d1_8c28_49c4e66d0b19.slice/crio-conmon-37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.618612 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.629472 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.629645 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.629669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.629704 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: E1004 03:42:57.629715 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 03:42:58.129695234 +0000 UTC m=+152.304318447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.647203 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.715323 4726 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T03:42:56.9575706Z","Handler":null,"Name":""} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.725882 4726 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.726202 4726 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.730704 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.736486 4726 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.736518 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.751353 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.763647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-gq29w\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.831402 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.840847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.937228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.947559 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:57 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:57 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:57 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.947647 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:57 crc kubenswrapper[4726]: W1004 03:42:57.952225 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod41ad0732_1ff2_4baf_a26e_80ede56b4a30.slice/crio-e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863 WatchSource:0}: Error finding container e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863: Status 404 returned error can't find the container with id e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.971524 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"41ad0732-1ff2-4baf-a26e-80ede56b4a30","Type":"ContainerStarted","Data":"e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.972765 4726 generic.go:334] "Generic (PLEG): container finished" podID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerID="3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed" exitCode=0 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.972840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerDied","Data":"3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.972901 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerStarted","Data":"d457e3d7c62d68d3d2b5a65761d490be51139beea124181183ff7a5dc944d26a"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.974933 4726 generic.go:334] "Generic (PLEG): container finished" podID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerID="37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa" exitCode=0 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.975001 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerDied","Data":"37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.975027 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerStarted","Data":"e3de63caa8dabdc349f7e96c808ccaf92b88bd948764797c7e9266dfc46546ba"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.976313 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerID="a43b2ae672193177047e9dfbed80a4ddbe7573d65c09a6ec2e1e9567c4b9da9e" exitCode=0 Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.976346 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerDied","Data":"a43b2ae672193177047e9dfbed80a4ddbe7573d65c09a6ec2e1e9567c4b9da9e"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.976372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerStarted","Data":"f497f2f5cbd9de7024a2ab0118f4544b32bde1ddf6fff1d6534343d6e130a6c6"} Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.976905 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:42:57 crc kubenswrapper[4726]: I1004 03:42:57.981778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" event={"ID":"72ca42a7-1fe7-4df8-b877-d0d5f94291fe","Type":"ContainerStarted","Data":"ee71ea04a1a3ddd355c5a3f3cd26c4bf114d190057b48d44ac688995dd62ad92"} Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.016193 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" event={"ID":"1d249e87-e695-4058-8b58-a068fcd85857","Type":"ContainerDied","Data":"10cfd0366c7d1692c742e3bfbbf050de003a8daa1ed47032313dfa8bdc81e4f1"} Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.016230 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10cfd0366c7d1692c742e3bfbbf050de003a8daa1ed47032313dfa8bdc81e4f1" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.016226 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.017423 4726 generic.go:334] "Generic (PLEG): container finished" podID="e26f8972-922a-46a9-859f-1b388b4fa660" containerID="2689735a37aa419165f42b7024b33b4574eabdf4a25ac1e3d80afa2e397a7fc8" exitCode=0 Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.017586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerDied","Data":"2689735a37aa419165f42b7024b33b4574eabdf4a25ac1e3d80afa2e397a7fc8"} Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.017629 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerStarted","Data":"7678d656d4797b406cced4de35f188bfa0fd7a22bb26d81845e7944686a90ded"} Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.035733 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.061368 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9l5gq" podStartSLOduration=12.061349753 podStartE2EDuration="12.061349753s" podCreationTimestamp="2025-10-04 03:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:58.057747554 +0000 UTC m=+152.232370757" watchObservedRunningTime="2025-10-04 03:42:58.061349753 +0000 UTC m=+152.235972966" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.133201 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.134129 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.138402 4726 patch_prober.go:28] interesting pod/console-f9d7485db-x6s4v container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.138438 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-x6s4v" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.202534 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.202587 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.213035 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.220991 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.222356 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.222402 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.222523 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.222565 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.226885 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:42:58 crc kubenswrapper[4726]: W1004 03:42:58.233223 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda88cfddf_a8f8_46a9_b581_7993f7d85ef4.slice/crio-248fca86822452501376586253a7fe9b13f5e664426ddeb5eb8e416bcea8c763 WatchSource:0}: Error finding container 248fca86822452501376586253a7fe9b13f5e664426ddeb5eb8e416bcea8c763: Status 404 returned error can't find the container with id 248fca86822452501376586253a7fe9b13f5e664426ddeb5eb8e416bcea8c763 Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.266342 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wp2cj" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.481725 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.509745 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.581602 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.582814 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.587309 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.607589 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.741722 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.741772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.741868 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc6lw\" (UniqueName: \"kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.836285 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.838475 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.843078 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.843134 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc6lw\" (UniqueName: \"kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.843235 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.843661 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.844375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.887924 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc6lw\" (UniqueName: \"kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw\") pod \"redhat-marketplace-t2vjb\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.902958 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.944164 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.946761 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:58 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:58 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:58 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.946818 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.979257 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.981868 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:58 crc kubenswrapper[4726]: I1004 03:42:58.988558 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.034776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" event={"ID":"a88cfddf-a8f8-46a9-b581-7993f7d85ef4","Type":"ContainerStarted","Data":"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60"} Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.034815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" event={"ID":"a88cfddf-a8f8-46a9-b581-7993f7d85ef4","Type":"ContainerStarted","Data":"248fca86822452501376586253a7fe9b13f5e664426ddeb5eb8e416bcea8c763"} Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.035608 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.041086 4726 generic.go:334] "Generic (PLEG): container finished" podID="41ad0732-1ff2-4baf-a26e-80ede56b4a30" containerID="9fa91f5952c7b0cedd7e463cfed959485d30402fd7027a2ac7bc186594dbd401" exitCode=0 Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.041517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"41ad0732-1ff2-4baf-a26e-80ede56b4a30","Type":"ContainerDied","Data":"9fa91f5952c7b0cedd7e463cfed959485d30402fd7027a2ac7bc186594dbd401"} Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.049686 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f7l7q" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.055833 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" podStartSLOduration=131.055823487 podStartE2EDuration="2m11.055823487s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:42:59.054979234 +0000 UTC m=+153.229602447" watchObservedRunningTime="2025-10-04 03:42:59.055823487 +0000 UTC m=+153.230446690" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.072697 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-w6fz4" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.102154 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.106642 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.114564 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.114603 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.115891 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.116489 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.137371 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6d56h" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.146691 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.146748 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.146859 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn5ld\" (UniqueName: \"kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.248534 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn5ld\" (UniqueName: \"kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.248603 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.248698 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.248720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.248751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.249285 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.249327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.273413 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn5ld\" (UniqueName: \"kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld\") pod \"redhat-marketplace-lm5dj\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.306047 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.350274 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.350381 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.350456 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.373586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.415198 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.459189 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.568789 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.588481 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.589558 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.601259 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.601681 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.663034 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.663239 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.663293 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fmtr\" (UniqueName: \"kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.764569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.764606 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fmtr\" (UniqueName: \"kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.764639 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.765061 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.766053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.792263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fmtr\" (UniqueName: \"kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr\") pod \"redhat-operators-lvtrq\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.899663 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.929331 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:42:59 crc kubenswrapper[4726]: W1004 03:42:59.941856 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda066b440_738a_4102_9c19_baa1e92403a3.slice/crio-d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae WatchSource:0}: Error finding container d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae: Status 404 returned error can't find the container with id d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.946627 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:42:59 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:42:59 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:42:59 crc kubenswrapper[4726]: healthz check failed Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.946669 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.983579 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.984914 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:42:59 crc kubenswrapper[4726]: I1004 03:42:59.993489 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.060488 4726 generic.go:334] "Generic (PLEG): container finished" podID="97de0531-09f7-4b57-b11c-53576c47ba04" containerID="6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d" exitCode=0 Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.060577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerDied","Data":"6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d"} Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.060711 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerStarted","Data":"b2cb56f7559b73657b60a193d61dfd19bbd1cee82f7c8d18bb109f435b94cbcc"} Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.067458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a066b440-738a-4102-9c19-baa1e92403a3","Type":"ContainerStarted","Data":"d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae"} Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.068197 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx28t\" (UniqueName: \"kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.068235 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.068257 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.087021 4726 generic.go:334] "Generic (PLEG): container finished" podID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerID="9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f" exitCode=0 Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.087263 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerDied","Data":"9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f"} Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.087315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerStarted","Data":"d13100b2226849a997b1708b8d59f6bfea1a82af76aec090d9420f360aac494d"} Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.169943 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx28t\" (UniqueName: \"kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.171491 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.171915 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.189827 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.190835 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.194809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx28t\" (UniqueName: \"kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t\") pod \"redhat-operators-z2nhb\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.310131 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.315254 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.374349 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir\") pod \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.374407 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access\") pod \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\" (UID: \"41ad0732-1ff2-4baf-a26e-80ede56b4a30\") " Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.374569 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "41ad0732-1ff2-4baf-a26e-80ede56b4a30" (UID: "41ad0732-1ff2-4baf-a26e-80ede56b4a30"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.375046 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.378510 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "41ad0732-1ff2-4baf-a26e-80ede56b4a30" (UID: "41ad0732-1ff2-4baf-a26e-80ede56b4a30"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.445914 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.476588 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41ad0732-1ff2-4baf-a26e-80ede56b4a30-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:43:00 crc kubenswrapper[4726]: W1004 03:43:00.505906 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod318cd44c_dd4c_4859_ad8b_3f4f308771d0.slice/crio-6118a21fae8d6e2b84ddd03dc4d46e1781cc768672387cc35d0817f704205b73 WatchSource:0}: Error finding container 6118a21fae8d6e2b84ddd03dc4d46e1781cc768672387cc35d0817f704205b73: Status 404 returned error can't find the container with id 6118a21fae8d6e2b84ddd03dc4d46e1781cc768672387cc35d0817f704205b73 Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.557575 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:43:00 crc kubenswrapper[4726]: W1004 03:43:00.607341 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod627806ff_dbd1_4cc3_9dfd_1bab9db62b85.slice/crio-49b1a62fd18bd2a59033ec0ddc2ccc69c3ad024d952520922d1a7f5c8225c88f WatchSource:0}: Error finding container 49b1a62fd18bd2a59033ec0ddc2ccc69c3ad024d952520922d1a7f5c8225c88f: Status 404 returned error can't find the container with id 49b1a62fd18bd2a59033ec0ddc2ccc69c3ad024d952520922d1a7f5c8225c88f Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.947028 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:43:00 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:43:00 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:43:00 crc kubenswrapper[4726]: healthz check failed Oct 04 03:43:00 crc kubenswrapper[4726]: I1004 03:43:00.947076 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.098845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a066b440-738a-4102-9c19-baa1e92403a3","Type":"ContainerStarted","Data":"ae9105e115dac48f9a7ff4a8bd093c46828f184f4a63169f9bdcca2c79d9c1c8"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.120921 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.120899988 podStartE2EDuration="2.120899988s" podCreationTimestamp="2025-10-04 03:42:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:43:01.110705596 +0000 UTC m=+155.285328809" watchObservedRunningTime="2025-10-04 03:43:01.120899988 +0000 UTC m=+155.295523201" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.124710 4726 generic.go:334] "Generic (PLEG): container finished" podID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerID="85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70" exitCode=0 Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.124789 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerDied","Data":"85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.124815 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerStarted","Data":"6118a21fae8d6e2b84ddd03dc4d46e1781cc768672387cc35d0817f704205b73"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.126872 4726 generic.go:334] "Generic (PLEG): container finished" podID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerID="63d5c32438a601bdb513a8a1146196b9d19a1433a17458e7dc7345eb7e39d1a9" exitCode=0 Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.126954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerDied","Data":"63d5c32438a601bdb513a8a1146196b9d19a1433a17458e7dc7345eb7e39d1a9"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.126990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerStarted","Data":"49b1a62fd18bd2a59033ec0ddc2ccc69c3ad024d952520922d1a7f5c8225c88f"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.129704 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"41ad0732-1ff2-4baf-a26e-80ede56b4a30","Type":"ContainerDied","Data":"e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863"} Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.129740 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e1d79f992cc4dc1e04f91d0324c9a452a82ee7a558a781c2020332bb3ef863" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.129740 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.572605 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.576598 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2s6pg" Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.946382 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:43:01 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:43:01 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:43:01 crc kubenswrapper[4726]: healthz check failed Oct 04 03:43:01 crc kubenswrapper[4726]: I1004 03:43:01.946433 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:43:02 crc kubenswrapper[4726]: I1004 03:43:02.150360 4726 generic.go:334] "Generic (PLEG): container finished" podID="a066b440-738a-4102-9c19-baa1e92403a3" containerID="ae9105e115dac48f9a7ff4a8bd093c46828f184f4a63169f9bdcca2c79d9c1c8" exitCode=0 Oct 04 03:43:02 crc kubenswrapper[4726]: I1004 03:43:02.150445 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a066b440-738a-4102-9c19-baa1e92403a3","Type":"ContainerDied","Data":"ae9105e115dac48f9a7ff4a8bd093c46828f184f4a63169f9bdcca2c79d9c1c8"} Oct 04 03:43:02 crc kubenswrapper[4726]: I1004 03:43:02.946356 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:43:02 crc kubenswrapper[4726]: [-]has-synced failed: reason withheld Oct 04 03:43:02 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:43:02 crc kubenswrapper[4726]: healthz check failed Oct 04 03:43:02 crc kubenswrapper[4726]: I1004 03:43:02.946411 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.462944 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.525478 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access\") pod \"a066b440-738a-4102-9c19-baa1e92403a3\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.531288 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir\") pod \"a066b440-738a-4102-9c19-baa1e92403a3\" (UID: \"a066b440-738a-4102-9c19-baa1e92403a3\") " Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.531678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a066b440-738a-4102-9c19-baa1e92403a3" (UID: "a066b440-738a-4102-9c19-baa1e92403a3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.538536 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a066b440-738a-4102-9c19-baa1e92403a3" (UID: "a066b440-738a-4102-9c19-baa1e92403a3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.632888 4726 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a066b440-738a-4102-9c19-baa1e92403a3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.632925 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a066b440-738a-4102-9c19-baa1e92403a3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.946629 4726 patch_prober.go:28] interesting pod/router-default-5444994796-zx9sv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 03:43:03 crc kubenswrapper[4726]: [+]has-synced ok Oct 04 03:43:03 crc kubenswrapper[4726]: [+]process-running ok Oct 04 03:43:03 crc kubenswrapper[4726]: healthz check failed Oct 04 03:43:03 crc kubenswrapper[4726]: I1004 03:43:03.946716 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zx9sv" podUID="42d8eec6-1736-44e1-adfd-995ecbde57a5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.173988 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a066b440-738a-4102-9c19-baa1e92403a3","Type":"ContainerDied","Data":"d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae"} Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.174030 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0ae09ed132a87f016127aa98942aafb2039ae3c65f3dac0621844d844d1e7ae" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.174036 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.180863 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-28vhf" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.187823 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.187869 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.946609 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:43:04 crc kubenswrapper[4726]: I1004 03:43:04.948874 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zx9sv" Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.139863 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.152041 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.222972 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.223075 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.230832 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:08 crc kubenswrapper[4726]: I1004 03:43:08.230875 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:10 crc kubenswrapper[4726]: I1004 03:43:10.839529 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:43:10 crc kubenswrapper[4726]: I1004 03:43:10.851183 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/77b969b7-0698-4589-a6cf-c08cc779ffe2-metrics-certs\") pod \"network-metrics-daemon-9qn78\" (UID: \"77b969b7-0698-4589-a6cf-c08cc779ffe2\") " pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:43:10 crc kubenswrapper[4726]: I1004 03:43:10.857821 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9qn78" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.053335 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.222389 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.222448 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.222491 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.222500 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.222566 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.223048 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"158844b1e06cfd5e9a24e38f604cac481dd2a469d7017774464add6666343d4e"} pod="openshift-console/downloads-7954f5f757-nw2zq" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.223164 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" containerID="cri-o://158844b1e06cfd5e9a24e38f604cac481dd2a469d7017774464add6666343d4e" gracePeriod=2 Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.223643 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:18 crc kubenswrapper[4726]: I1004 03:43:18.223726 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:19 crc kubenswrapper[4726]: I1004 03:43:19.278664 4726 generic.go:334] "Generic (PLEG): container finished" podID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerID="158844b1e06cfd5e9a24e38f604cac481dd2a469d7017774464add6666343d4e" exitCode=0 Oct 04 03:43:19 crc kubenswrapper[4726]: I1004 03:43:19.278749 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nw2zq" event={"ID":"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d","Type":"ContainerDied","Data":"158844b1e06cfd5e9a24e38f604cac481dd2a469d7017774464add6666343d4e"} Oct 04 03:43:28 crc kubenswrapper[4726]: I1004 03:43:28.223307 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:28 crc kubenswrapper[4726]: I1004 03:43:28.224173 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:29 crc kubenswrapper[4726]: I1004 03:43:29.087301 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-nw97k" Oct 04 03:43:34 crc kubenswrapper[4726]: I1004 03:43:34.188397 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:43:34 crc kubenswrapper[4726]: I1004 03:43:34.188939 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:43:35 crc kubenswrapper[4726]: I1004 03:43:35.574686 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 03:43:38 crc kubenswrapper[4726]: I1004 03:43:38.224028 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:38 crc kubenswrapper[4726]: I1004 03:43:38.224144 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:42 crc kubenswrapper[4726]: E1004 03:43:42.623832 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:43:42 crc kubenswrapper[4726]: E1004 03:43:42.624315 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bx28t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-z2nhb_openshift-marketplace(627806ff-dbd1-4cc3-9dfd-1bab9db62b85): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:42 crc kubenswrapper[4726]: E1004 03:43:42.625499 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-z2nhb" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" Oct 04 03:43:43 crc kubenswrapper[4726]: E1004 03:43:43.311270 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 03:43:43 crc kubenswrapper[4726]: E1004 03:43:43.311449 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8fmtr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lvtrq_openshift-marketplace(318cd44c-dd4c-4859-ad8b-3f4f308771d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:43 crc kubenswrapper[4726]: E1004 03:43:43.312668 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lvtrq" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" Oct 04 03:43:46 crc kubenswrapper[4726]: E1004 03:43:46.782501 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lvtrq" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" Oct 04 03:43:46 crc kubenswrapper[4726]: E1004 03:43:46.782943 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-z2nhb" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" Oct 04 03:43:48 crc kubenswrapper[4726]: E1004 03:43:48.046811 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:43:48 crc kubenswrapper[4726]: E1004 03:43:48.047269 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxtvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-khfd7_openshift-marketplace(c8435186-bbfc-4907-bf0a-2a6b88613f07): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:48 crc kubenswrapper[4726]: E1004 03:43:48.048978 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-khfd7" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" Oct 04 03:43:48 crc kubenswrapper[4726]: I1004 03:43:48.222886 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:48 crc kubenswrapper[4726]: I1004 03:43:48.223304 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.301773 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-khfd7" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.367040 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.367232 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvthp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-vm256_openshift-marketplace(e26f8972-922a-46a9-859f-1b388b4fa660): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.368543 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-vm256" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.410021 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.410289 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cwr4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qzrv5_openshift-marketplace(e9ba01a7-5230-4c26-89e4-19cf823c3ded): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.411705 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qzrv5" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.602187 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-vm256" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.602197 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qzrv5" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.715762 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.715941 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ktbbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-swwmz_openshift-marketplace(747e7096-4e73-42d1-8c28-49c4e66d0b19): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:49 crc kubenswrapper[4726]: E1004 03:43:49.718097 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-swwmz" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.116497 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-swwmz" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" Oct 04 03:43:52 crc kubenswrapper[4726]: I1004 03:43:52.550609 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9qn78"] Oct 04 03:43:52 crc kubenswrapper[4726]: W1004 03:43:52.833674 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77b969b7_0698_4589_a6cf_c08cc779ffe2.slice/crio-3289a1df89830621947db7ca6abb82896700e77934e21a46ee3771f7fe2c3148 WatchSource:0}: Error finding container 3289a1df89830621947db7ca6abb82896700e77934e21a46ee3771f7fe2c3148: Status 404 returned error can't find the container with id 3289a1df89830621947db7ca6abb82896700e77934e21a46ee3771f7fe2c3148 Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.889711 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.889957 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jn5ld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lm5dj_openshift-marketplace(c65adfe0-81c6-4e17-be2a-9feff5929860): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.891762 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lm5dj" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.912795 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.912956 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nc6lw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-t2vjb_openshift-marketplace(97de0531-09f7-4b57-b11c-53576c47ba04): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:43:52 crc kubenswrapper[4726]: E1004 03:43:52.914065 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-t2vjb" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.471526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9qn78" event={"ID":"77b969b7-0698-4589-a6cf-c08cc779ffe2","Type":"ContainerStarted","Data":"a49e981bcf4473eba83fcb8226e5e2cb13eec099797cb6727ea686e803aaf496"} Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.471971 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9qn78" event={"ID":"77b969b7-0698-4589-a6cf-c08cc779ffe2","Type":"ContainerStarted","Data":"3068f9bfd3d186053ff71267ab5e0e1fae7573eaab12cc8b3ff6b9920cef1f41"} Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.471997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9qn78" event={"ID":"77b969b7-0698-4589-a6cf-c08cc779ffe2","Type":"ContainerStarted","Data":"3289a1df89830621947db7ca6abb82896700e77934e21a46ee3771f7fe2c3148"} Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.473955 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nw2zq" event={"ID":"ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d","Type":"ContainerStarted","Data":"c5476ff44ff4add004c461ff8f602221eac127858d540ddca8f9ddc5330f8e1e"} Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.474706 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.476700 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.476762 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:53 crc kubenswrapper[4726]: E1004 03:43:53.477061 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-t2vjb" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" Oct 04 03:43:53 crc kubenswrapper[4726]: E1004 03:43:53.477424 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lm5dj" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" Oct 04 03:43:53 crc kubenswrapper[4726]: I1004 03:43:53.497325 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9qn78" podStartSLOduration=185.497295513 podStartE2EDuration="3m5.497295513s" podCreationTimestamp="2025-10-04 03:40:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:43:53.495487744 +0000 UTC m=+207.670110997" watchObservedRunningTime="2025-10-04 03:43:53.497295513 +0000 UTC m=+207.671918756" Oct 04 03:43:54 crc kubenswrapper[4726]: I1004 03:43:54.483623 4726 patch_prober.go:28] interesting pod/downloads-7954f5f757-nw2zq container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 04 03:43:54 crc kubenswrapper[4726]: I1004 03:43:54.483713 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nw2zq" podUID="ed1d5fb4-ef43-4fca-9a16-4f67b8537c0d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 04 03:43:58 crc kubenswrapper[4726]: I1004 03:43:58.241841 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-nw2zq" Oct 04 03:44:04 crc kubenswrapper[4726]: I1004 03:44:04.188612 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:44:04 crc kubenswrapper[4726]: I1004 03:44:04.190390 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:44:04 crc kubenswrapper[4726]: I1004 03:44:04.190468 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:44:04 crc kubenswrapper[4726]: I1004 03:44:04.191393 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:44:04 crc kubenswrapper[4726]: I1004 03:44:04.191506 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc" gracePeriod=600 Oct 04 03:44:05 crc kubenswrapper[4726]: I1004 03:44:05.563438 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc" exitCode=0 Oct 04 03:44:05 crc kubenswrapper[4726]: I1004 03:44:05.563487 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.601469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerStarted","Data":"10dfa947b0572e6ced2edbc3d99e0d9a89a8e20a05c3fcf9a5ae99e006ce5e1c"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.603036 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerStarted","Data":"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.605518 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerStarted","Data":"9b3307058bcd829ec6067240f43ff8a364b9b132b12ffb325ed40749ab87929c"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.607345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerStarted","Data":"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.609459 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerStarted","Data":"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.611755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e"} Oct 04 03:44:09 crc kubenswrapper[4726]: I1004 03:44:09.613795 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerStarted","Data":"5ef3f35871c8cccd36dd8596999624c9139c40abf3f5a19fb083c8bd46c584fb"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.624413 4726 generic.go:334] "Generic (PLEG): container finished" podID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerID="e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.624515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerDied","Data":"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.629428 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerID="9b3307058bcd829ec6067240f43ff8a364b9b132b12ffb325ed40749ab87929c" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.629538 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerDied","Data":"9b3307058bcd829ec6067240f43ff8a364b9b132b12ffb325ed40749ab87929c"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.637080 4726 generic.go:334] "Generic (PLEG): container finished" podID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerID="58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.637162 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerDied","Data":"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.641939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerStarted","Data":"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.645759 4726 generic.go:334] "Generic (PLEG): container finished" podID="97de0531-09f7-4b57-b11c-53576c47ba04" containerID="d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.645862 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerDied","Data":"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.647985 4726 generic.go:334] "Generic (PLEG): container finished" podID="e26f8972-922a-46a9-859f-1b388b4fa660" containerID="5ef3f35871c8cccd36dd8596999624c9139c40abf3f5a19fb083c8bd46c584fb" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.648079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerDied","Data":"5ef3f35871c8cccd36dd8596999624c9139c40abf3f5a19fb083c8bd46c584fb"} Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.652249 4726 generic.go:334] "Generic (PLEG): container finished" podID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerID="10dfa947b0572e6ced2edbc3d99e0d9a89a8e20a05c3fcf9a5ae99e006ce5e1c" exitCode=0 Oct 04 03:44:10 crc kubenswrapper[4726]: I1004 03:44:10.652314 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerDied","Data":"10dfa947b0572e6ced2edbc3d99e0d9a89a8e20a05c3fcf9a5ae99e006ce5e1c"} Oct 04 03:44:11 crc kubenswrapper[4726]: I1004 03:44:11.662630 4726 generic.go:334] "Generic (PLEG): container finished" podID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerID="de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d" exitCode=0 Oct 04 03:44:11 crc kubenswrapper[4726]: I1004 03:44:11.662816 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerDied","Data":"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d"} Oct 04 03:44:11 crc kubenswrapper[4726]: I1004 03:44:11.667446 4726 generic.go:334] "Generic (PLEG): container finished" podID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerID="e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65" exitCode=0 Oct 04 03:44:11 crc kubenswrapper[4726]: I1004 03:44:11.667532 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerDied","Data":"e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65"} Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.679740 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerStarted","Data":"fc48f84a9795a5540a2478901648a946dd2c5f3c37dec4b3f1dcb78d44ed51e2"} Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.686827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerStarted","Data":"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753"} Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.691996 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerStarted","Data":"20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4"} Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.711649 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qzrv5" podStartSLOduration=2.965790866 podStartE2EDuration="1m16.711624951s" podCreationTimestamp="2025-10-04 03:42:56 +0000 UTC" firstStartedPulling="2025-10-04 03:42:57.979223694 +0000 UTC m=+152.153846907" lastFinishedPulling="2025-10-04 03:44:11.725057749 +0000 UTC m=+225.899680992" observedRunningTime="2025-10-04 03:44:12.706343605 +0000 UTC m=+226.880966848" watchObservedRunningTime="2025-10-04 03:44:12.711624951 +0000 UTC m=+226.886248174" Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.733552 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2nhb" podStartSLOduration=2.67937332 podStartE2EDuration="1m13.733524143s" podCreationTimestamp="2025-10-04 03:42:59 +0000 UTC" firstStartedPulling="2025-10-04 03:43:01.128777616 +0000 UTC m=+155.303400829" lastFinishedPulling="2025-10-04 03:44:12.182928409 +0000 UTC m=+226.357551652" observedRunningTime="2025-10-04 03:44:12.73158684 +0000 UTC m=+226.906210093" watchObservedRunningTime="2025-10-04 03:44:12.733524143 +0000 UTC m=+226.908147396" Oct 04 03:44:12 crc kubenswrapper[4726]: I1004 03:44:12.759364 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lvtrq" podStartSLOduration=3.29907113 podStartE2EDuration="1m13.759339565s" podCreationTimestamp="2025-10-04 03:42:59 +0000 UTC" firstStartedPulling="2025-10-04 03:43:01.127710146 +0000 UTC m=+155.302333359" lastFinishedPulling="2025-10-04 03:44:11.587978541 +0000 UTC m=+225.762601794" observedRunningTime="2025-10-04 03:44:12.75226306 +0000 UTC m=+226.926886323" watchObservedRunningTime="2025-10-04 03:44:12.759339565 +0000 UTC m=+226.933962818" Oct 04 03:44:13 crc kubenswrapper[4726]: I1004 03:44:13.705512 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerStarted","Data":"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70"} Oct 04 03:44:14 crc kubenswrapper[4726]: I1004 03:44:14.737255 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-khfd7" podStartSLOduration=3.580494427 podStartE2EDuration="1m18.737235509s" podCreationTimestamp="2025-10-04 03:42:56 +0000 UTC" firstStartedPulling="2025-10-04 03:42:57.976641532 +0000 UTC m=+152.151264745" lastFinishedPulling="2025-10-04 03:44:13.133382604 +0000 UTC m=+227.308005827" observedRunningTime="2025-10-04 03:44:14.736425927 +0000 UTC m=+228.911049140" watchObservedRunningTime="2025-10-04 03:44:14.737235509 +0000 UTC m=+228.911858722" Oct 04 03:44:16 crc kubenswrapper[4726]: I1004 03:44:16.720422 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:44:16 crc kubenswrapper[4726]: I1004 03:44:16.720462 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.114019 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.114396 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.509196 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.519919 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.729634 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerStarted","Data":"cd1d13e43269c88138854a94a75a389c2beff829ba6a1e78f913c24b70da56bd"} Oct 04 03:44:17 crc kubenswrapper[4726]: I1004 03:44:17.862726 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:18 crc kubenswrapper[4726]: I1004 03:44:18.759043 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vm256" podStartSLOduration=4.067242046 podStartE2EDuration="1m22.759026206s" podCreationTimestamp="2025-10-04 03:42:56 +0000 UTC" firstStartedPulling="2025-10-04 03:42:58.02035346 +0000 UTC m=+152.194976673" lastFinishedPulling="2025-10-04 03:44:16.71213762 +0000 UTC m=+230.886760833" observedRunningTime="2025-10-04 03:44:18.756825665 +0000 UTC m=+232.931448878" watchObservedRunningTime="2025-10-04 03:44:18.759026206 +0000 UTC m=+232.933649419" Oct 04 03:44:18 crc kubenswrapper[4726]: I1004 03:44:18.836162 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:44:19 crc kubenswrapper[4726]: I1004 03:44:19.743147 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qzrv5" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="registry-server" containerID="cri-o://fc48f84a9795a5540a2478901648a946dd2c5f3c37dec4b3f1dcb78d44ed51e2" gracePeriod=2 Oct 04 03:44:19 crc kubenswrapper[4726]: I1004 03:44:19.932002 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:44:19 crc kubenswrapper[4726]: I1004 03:44:19.932057 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:44:19 crc kubenswrapper[4726]: I1004 03:44:19.990260 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:44:20 crc kubenswrapper[4726]: I1004 03:44:20.310443 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:20 crc kubenswrapper[4726]: I1004 03:44:20.310854 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:20 crc kubenswrapper[4726]: I1004 03:44:20.382207 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:20 crc kubenswrapper[4726]: I1004 03:44:20.804913 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:44:20 crc kubenswrapper[4726]: I1004 03:44:20.817028 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:21 crc kubenswrapper[4726]: I1004 03:44:21.778136 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerStarted","Data":"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9"} Oct 04 03:44:21 crc kubenswrapper[4726]: I1004 03:44:21.781827 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerID="fc48f84a9795a5540a2478901648a946dd2c5f3c37dec4b3f1dcb78d44ed51e2" exitCode=0 Oct 04 03:44:21 crc kubenswrapper[4726]: I1004 03:44:21.782995 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerDied","Data":"fc48f84a9795a5540a2478901648a946dd2c5f3c37dec4b3f1dcb78d44ed51e2"} Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.244787 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.387323 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.586422 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwr4g\" (UniqueName: \"kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g\") pod \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.586498 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content\") pod \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.586564 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities\") pod \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\" (UID: \"e9ba01a7-5230-4c26-89e4-19cf823c3ded\") " Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.588189 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities" (OuterVolumeSpecName: "utilities") pod "e9ba01a7-5230-4c26-89e4-19cf823c3ded" (UID: "e9ba01a7-5230-4c26-89e4-19cf823c3ded"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.607423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g" (OuterVolumeSpecName: "kube-api-access-cwr4g") pod "e9ba01a7-5230-4c26-89e4-19cf823c3ded" (UID: "e9ba01a7-5230-4c26-89e4-19cf823c3ded"). InnerVolumeSpecName "kube-api-access-cwr4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.687947 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwr4g\" (UniqueName: \"kubernetes.io/projected/e9ba01a7-5230-4c26-89e4-19cf823c3ded-kube-api-access-cwr4g\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.688008 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.793093 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2nhb" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="registry-server" containerID="cri-o://20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" gracePeriod=2 Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.793913 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qzrv5" Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.794226 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qzrv5" event={"ID":"e9ba01a7-5230-4c26-89e4-19cf823c3ded","Type":"ContainerDied","Data":"f497f2f5cbd9de7024a2ab0118f4544b32bde1ddf6fff1d6534343d6e130a6c6"} Oct 04 03:44:22 crc kubenswrapper[4726]: I1004 03:44:22.794295 4726 scope.go:117] "RemoveContainer" containerID="fc48f84a9795a5540a2478901648a946dd2c5f3c37dec4b3f1dcb78d44ed51e2" Oct 04 03:44:23 crc kubenswrapper[4726]: I1004 03:44:23.831721 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t2vjb" podStartSLOduration=5.75973952 podStartE2EDuration="1m25.831693546s" podCreationTimestamp="2025-10-04 03:42:58 +0000 UTC" firstStartedPulling="2025-10-04 03:43:00.067368432 +0000 UTC m=+154.241991645" lastFinishedPulling="2025-10-04 03:44:20.139322448 +0000 UTC m=+234.313945671" observedRunningTime="2025-10-04 03:44:23.825261419 +0000 UTC m=+237.999884672" watchObservedRunningTime="2025-10-04 03:44:23.831693546 +0000 UTC m=+238.006316759" Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.000646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9ba01a7-5230-4c26-89e4-19cf823c3ded" (UID: "e9ba01a7-5230-4c26-89e4-19cf823c3ded"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.044631 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9ba01a7-5230-4c26-89e4-19cf823c3ded-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.141770 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.148886 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qzrv5"] Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.520776 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" path="/var/lib/kubelet/pods/e9ba01a7-5230-4c26-89e4-19cf823c3ded/volumes" Oct 04 03:44:26 crc kubenswrapper[4726]: I1004 03:44:26.793545 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.352547 4726 scope.go:117] "RemoveContainer" containerID="9b3307058bcd829ec6067240f43ff8a364b9b132b12ffb325ed40749ab87929c" Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.395149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.396880 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.448452 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.834441 4726 generic.go:334] "Generic (PLEG): container finished" podID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerID="20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" exitCode=0 Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.834537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerDied","Data":"20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4"} Oct 04 03:44:27 crc kubenswrapper[4726]: I1004 03:44:27.900866 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:28 crc kubenswrapper[4726]: I1004 03:44:28.559496 4726 scope.go:117] "RemoveContainer" containerID="a43b2ae672193177047e9dfbed80a4ddbe7573d65c09a6ec2e1e9567c4b9da9e" Oct 04 03:44:28 crc kubenswrapper[4726]: I1004 03:44:28.904410 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:44:28 crc kubenswrapper[4726]: I1004 03:44:28.904495 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:44:28 crc kubenswrapper[4726]: I1004 03:44:28.973211 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:44:29 crc kubenswrapper[4726]: I1004 03:44:29.249885 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:44:29 crc kubenswrapper[4726]: I1004 03:44:29.867796 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vm256" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="registry-server" containerID="cri-o://cd1d13e43269c88138854a94a75a389c2beff829ba6a1e78f913c24b70da56bd" gracePeriod=2 Oct 04 03:44:29 crc kubenswrapper[4726]: I1004 03:44:29.936348 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:44:30 crc kubenswrapper[4726]: E1004 03:44:30.312176 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4 is running failed: container process not found" containerID="20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:44:30 crc kubenswrapper[4726]: E1004 03:44:30.313303 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4 is running failed: container process not found" containerID="20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:44:30 crc kubenswrapper[4726]: E1004 03:44:30.313912 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4 is running failed: container process not found" containerID="20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 03:44:30 crc kubenswrapper[4726]: E1004 03:44:30.313983 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-z2nhb" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="registry-server" Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.872626 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.888315 4726 generic.go:334] "Generic (PLEG): container finished" podID="e26f8972-922a-46a9-859f-1b388b4fa660" containerID="cd1d13e43269c88138854a94a75a389c2beff829ba6a1e78f913c24b70da56bd" exitCode=0 Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.888979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerDied","Data":"cd1d13e43269c88138854a94a75a389c2beff829ba6a1e78f913c24b70da56bd"} Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.893182 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2nhb" event={"ID":"627806ff-dbd1-4cc3-9dfd-1bab9db62b85","Type":"ContainerDied","Data":"49b1a62fd18bd2a59033ec0ddc2ccc69c3ad024d952520922d1a7f5c8225c88f"} Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.893314 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2nhb" Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.955550 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities\") pod \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.955648 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx28t\" (UniqueName: \"kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t\") pod \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.955706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content\") pod \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\" (UID: \"627806ff-dbd1-4cc3-9dfd-1bab9db62b85\") " Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.956407 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities" (OuterVolumeSpecName: "utilities") pod "627806ff-dbd1-4cc3-9dfd-1bab9db62b85" (UID: "627806ff-dbd1-4cc3-9dfd-1bab9db62b85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:32 crc kubenswrapper[4726]: I1004 03:44:32.963999 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t" (OuterVolumeSpecName: "kube-api-access-bx28t") pod "627806ff-dbd1-4cc3-9dfd-1bab9db62b85" (UID: "627806ff-dbd1-4cc3-9dfd-1bab9db62b85"). InnerVolumeSpecName "kube-api-access-bx28t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.057844 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.058283 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx28t\" (UniqueName: \"kubernetes.io/projected/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-kube-api-access-bx28t\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.099815 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "627806ff-dbd1-4cc3-9dfd-1bab9db62b85" (UID: "627806ff-dbd1-4cc3-9dfd-1bab9db62b85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.160175 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627806ff-dbd1-4cc3-9dfd-1bab9db62b85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.171693 4726 scope.go:117] "RemoveContainer" containerID="20eb08d9d2296e7f825e03daed37cb10a8eedf0d16c381a2db61e952813f83b4" Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.244194 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:44:33 crc kubenswrapper[4726]: I1004 03:44:33.251101 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2nhb"] Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.089494 4726 scope.go:117] "RemoveContainer" containerID="10dfa947b0572e6ced2edbc3d99e0d9a89a8e20a05c3fcf9a5ae99e006ce5e1c" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.198254 4726 scope.go:117] "RemoveContainer" containerID="63d5c32438a601bdb513a8a1146196b9d19a1433a17458e7dc7345eb7e39d1a9" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.210686 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.272873 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities\") pod \"e26f8972-922a-46a9-859f-1b388b4fa660\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.272924 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content\") pod \"e26f8972-922a-46a9-859f-1b388b4fa660\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.272967 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvthp\" (UniqueName: \"kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp\") pod \"e26f8972-922a-46a9-859f-1b388b4fa660\" (UID: \"e26f8972-922a-46a9-859f-1b388b4fa660\") " Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.274327 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities" (OuterVolumeSpecName: "utilities") pod "e26f8972-922a-46a9-859f-1b388b4fa660" (UID: "e26f8972-922a-46a9-859f-1b388b4fa660"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.280404 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp" (OuterVolumeSpecName: "kube-api-access-rvthp") pod "e26f8972-922a-46a9-859f-1b388b4fa660" (UID: "e26f8972-922a-46a9-859f-1b388b4fa660"). InnerVolumeSpecName "kube-api-access-rvthp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.339341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e26f8972-922a-46a9-859f-1b388b4fa660" (UID: "e26f8972-922a-46a9-859f-1b388b4fa660"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.376443 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.376724 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e26f8972-922a-46a9-859f-1b388b4fa660-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.376837 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvthp\" (UniqueName: \"kubernetes.io/projected/e26f8972-922a-46a9-859f-1b388b4fa660-kube-api-access-rvthp\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.511764 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" path="/var/lib/kubelet/pods/627806ff-dbd1-4cc3-9dfd-1bab9db62b85/volumes" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.907763 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vm256" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.908001 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vm256" event={"ID":"e26f8972-922a-46a9-859f-1b388b4fa660","Type":"ContainerDied","Data":"7678d656d4797b406cced4de35f188bfa0fd7a22bb26d81845e7944686a90ded"} Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.908670 4726 scope.go:117] "RemoveContainer" containerID="cd1d13e43269c88138854a94a75a389c2beff829ba6a1e78f913c24b70da56bd" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.915605 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerStarted","Data":"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3"} Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.922425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerStarted","Data":"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d"} Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.942337 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.943898 4726 scope.go:117] "RemoveContainer" containerID="5ef3f35871c8cccd36dd8596999624c9139c40abf3f5a19fb083c8bd46c584fb" Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.949687 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vm256"] Oct 04 03:44:34 crc kubenswrapper[4726]: I1004 03:44:34.974883 4726 scope.go:117] "RemoveContainer" containerID="2689735a37aa419165f42b7024b33b4574eabdf4a25ac1e3d80afa2e397a7fc8" Oct 04 03:44:35 crc kubenswrapper[4726]: I1004 03:44:35.960193 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swwmz" podStartSLOduration=3.847341351 podStartE2EDuration="1m39.960167969s" podCreationTimestamp="2025-10-04 03:42:56 +0000 UTC" firstStartedPulling="2025-10-04 03:42:57.976658703 +0000 UTC m=+152.151281916" lastFinishedPulling="2025-10-04 03:44:34.089485321 +0000 UTC m=+248.264108534" observedRunningTime="2025-10-04 03:44:35.954797451 +0000 UTC m=+250.129420694" watchObservedRunningTime="2025-10-04 03:44:35.960167969 +0000 UTC m=+250.134791222" Oct 04 03:44:35 crc kubenswrapper[4726]: I1004 03:44:35.981279 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lm5dj" podStartSLOduration=4.085295131 podStartE2EDuration="1m37.98125582s" podCreationTimestamp="2025-10-04 03:42:58 +0000 UTC" firstStartedPulling="2025-10-04 03:43:00.089567296 +0000 UTC m=+154.264190509" lastFinishedPulling="2025-10-04 03:44:33.985527925 +0000 UTC m=+248.160151198" observedRunningTime="2025-10-04 03:44:35.977669121 +0000 UTC m=+250.152292384" watchObservedRunningTime="2025-10-04 03:44:35.98125582 +0000 UTC m=+250.155879043" Oct 04 03:44:36 crc kubenswrapper[4726]: I1004 03:44:36.507950 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" path="/var/lib/kubelet/pods/e26f8972-922a-46a9-859f-1b388b4fa660/volumes" Oct 04 03:44:36 crc kubenswrapper[4726]: I1004 03:44:36.921685 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:44:36 crc kubenswrapper[4726]: I1004 03:44:36.921751 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:44:37 crc kubenswrapper[4726]: I1004 03:44:37.994096 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-swwmz" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="registry-server" probeResult="failure" output=< Oct 04 03:44:37 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 03:44:37 crc kubenswrapper[4726]: > Oct 04 03:44:39 crc kubenswrapper[4726]: I1004 03:44:39.307549 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:39 crc kubenswrapper[4726]: I1004 03:44:39.309449 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:39 crc kubenswrapper[4726]: I1004 03:44:39.380359 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:40 crc kubenswrapper[4726]: I1004 03:44:40.026698 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:41 crc kubenswrapper[4726]: I1004 03:44:41.241729 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:44:41 crc kubenswrapper[4726]: I1004 03:44:41.972390 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lm5dj" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="registry-server" containerID="cri-o://0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d" gracePeriod=2 Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.393853 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.492708 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content\") pod \"c65adfe0-81c6-4e17-be2a-9feff5929860\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.492751 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities\") pod \"c65adfe0-81c6-4e17-be2a-9feff5929860\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.492806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn5ld\" (UniqueName: \"kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld\") pod \"c65adfe0-81c6-4e17-be2a-9feff5929860\" (UID: \"c65adfe0-81c6-4e17-be2a-9feff5929860\") " Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.494742 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities" (OuterVolumeSpecName: "utilities") pod "c65adfe0-81c6-4e17-be2a-9feff5929860" (UID: "c65adfe0-81c6-4e17-be2a-9feff5929860"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.500475 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld" (OuterVolumeSpecName: "kube-api-access-jn5ld") pod "c65adfe0-81c6-4e17-be2a-9feff5929860" (UID: "c65adfe0-81c6-4e17-be2a-9feff5929860"). InnerVolumeSpecName "kube-api-access-jn5ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.524046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c65adfe0-81c6-4e17-be2a-9feff5929860" (UID: "c65adfe0-81c6-4e17-be2a-9feff5929860"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.594036 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.594098 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c65adfe0-81c6-4e17-be2a-9feff5929860-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.594160 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn5ld\" (UniqueName: \"kubernetes.io/projected/c65adfe0-81c6-4e17-be2a-9feff5929860-kube-api-access-jn5ld\") on node \"crc\" DevicePath \"\"" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.983022 4726 generic.go:334] "Generic (PLEG): container finished" podID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerID="0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d" exitCode=0 Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.983133 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerDied","Data":"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d"} Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.983186 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm5dj" event={"ID":"c65adfe0-81c6-4e17-be2a-9feff5929860","Type":"ContainerDied","Data":"d13100b2226849a997b1708b8d59f6bfea1a82af76aec090d9420f360aac494d"} Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.983223 4726 scope.go:117] "RemoveContainer" containerID="0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d" Oct 04 03:44:42 crc kubenswrapper[4726]: I1004 03:44:42.983453 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm5dj" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.009649 4726 scope.go:117] "RemoveContainer" containerID="de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.031623 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.034158 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm5dj"] Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.068791 4726 scope.go:117] "RemoveContainer" containerID="9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.088963 4726 scope.go:117] "RemoveContainer" containerID="0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d" Oct 04 03:44:43 crc kubenswrapper[4726]: E1004 03:44:43.089365 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d\": container with ID starting with 0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d not found: ID does not exist" containerID="0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.089395 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d"} err="failed to get container status \"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d\": rpc error: code = NotFound desc = could not find container \"0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d\": container with ID starting with 0f079e30d57cca4adf53cecc2ba799a4b50f17f91e1efc76dd3428bdfb3ca82d not found: ID does not exist" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.089418 4726 scope.go:117] "RemoveContainer" containerID="de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d" Oct 04 03:44:43 crc kubenswrapper[4726]: E1004 03:44:43.090134 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d\": container with ID starting with de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d not found: ID does not exist" containerID="de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.090198 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d"} err="failed to get container status \"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d\": rpc error: code = NotFound desc = could not find container \"de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d\": container with ID starting with de7e892635e3dfd22fdfacfb93ad856d309e1cd7f919eaa6ffad5aeb0dde4b9d not found: ID does not exist" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.090262 4726 scope.go:117] "RemoveContainer" containerID="9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f" Oct 04 03:44:43 crc kubenswrapper[4726]: E1004 03:44:43.090785 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f\": container with ID starting with 9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f not found: ID does not exist" containerID="9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f" Oct 04 03:44:43 crc kubenswrapper[4726]: I1004 03:44:43.090824 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f"} err="failed to get container status \"9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f\": rpc error: code = NotFound desc = could not find container \"9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f\": container with ID starting with 9553ca5111732bc74b7ddf938a57b85debf7f7703cb2eaa09eb6d8ba9132528f not found: ID does not exist" Oct 04 03:44:44 crc kubenswrapper[4726]: I1004 03:44:44.512452 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" path="/var/lib/kubelet/pods/c65adfe0-81c6-4e17-be2a-9feff5929860/volumes" Oct 04 03:44:46 crc kubenswrapper[4726]: I1004 03:44:46.960493 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:44:47 crc kubenswrapper[4726]: I1004 03:44:47.003719 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.142620 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn"] Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146190 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146234 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146248 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146258 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146277 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146285 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146311 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146319 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146330 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146338 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146353 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146361 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146379 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146387 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146396 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a066b440-738a-4102-9c19-baa1e92403a3" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146404 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a066b440-738a-4102-9c19-baa1e92403a3" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146421 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146429 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146445 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146453 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146476 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146483 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146493 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41ad0732-1ff2-4baf-a26e-80ede56b4a30" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146501 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="41ad0732-1ff2-4baf-a26e-80ede56b4a30" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146517 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146529 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="extract-utilities" Oct 04 03:45:00 crc kubenswrapper[4726]: E1004 03:45:00.146539 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146547 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="extract-content" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146867 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ba01a7-5230-4c26-89e4-19cf823c3ded" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146885 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a066b440-738a-4102-9c19-baa1e92403a3" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146908 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c65adfe0-81c6-4e17-be2a-9feff5929860" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146933 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="627806ff-dbd1-4cc3-9dfd-1bab9db62b85" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146943 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="41ad0732-1ff2-4baf-a26e-80ede56b4a30" containerName="pruner" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.146958 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e26f8972-922a-46a9-859f-1b388b4fa660" containerName="registry-server" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.147709 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.153869 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.153984 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.169446 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn"] Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.216518 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.216579 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7l5\" (UniqueName: \"kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.216611 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.317871 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.318009 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7l5\" (UniqueName: \"kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.318074 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.319065 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.327671 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.335101 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7l5\" (UniqueName: \"kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5\") pod \"collect-profiles-29325825-df8cn\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.477745 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:00 crc kubenswrapper[4726]: I1004 03:45:00.681119 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn"] Oct 04 03:45:01 crc kubenswrapper[4726]: I1004 03:45:01.078948 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" event={"ID":"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf","Type":"ContainerStarted","Data":"bd005a81442dc0147dc57f59565cf94f9d07985fa1244cccd45fac5e3e332a23"} Oct 04 03:45:01 crc kubenswrapper[4726]: I1004 03:45:01.079315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" event={"ID":"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf","Type":"ContainerStarted","Data":"134abf40b0dd91df9e9b17fba4e0b96591533974747e9ed0ed511d608f1dd221"} Oct 04 03:45:01 crc kubenswrapper[4726]: I1004 03:45:01.093005 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" podStartSLOduration=1.092988007 podStartE2EDuration="1.092988007s" podCreationTimestamp="2025-10-04 03:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:45:01.092179245 +0000 UTC m=+275.266802458" watchObservedRunningTime="2025-10-04 03:45:01.092988007 +0000 UTC m=+275.267611210" Oct 04 03:45:02 crc kubenswrapper[4726]: I1004 03:45:02.086159 4726 generic.go:334] "Generic (PLEG): container finished" podID="1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" containerID="bd005a81442dc0147dc57f59565cf94f9d07985fa1244cccd45fac5e3e332a23" exitCode=0 Oct 04 03:45:02 crc kubenswrapper[4726]: I1004 03:45:02.086205 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" event={"ID":"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf","Type":"ContainerDied","Data":"bd005a81442dc0147dc57f59565cf94f9d07985fa1244cccd45fac5e3e332a23"} Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.421843 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.555641 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf7l5\" (UniqueName: \"kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5\") pod \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.555689 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume\") pod \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.555710 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume\") pod \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\" (UID: \"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf\") " Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.556537 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume" (OuterVolumeSpecName: "config-volume") pod "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" (UID: "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.564223 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" (UID: "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.565420 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5" (OuterVolumeSpecName: "kube-api-access-tf7l5") pod "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" (UID: "1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf"). InnerVolumeSpecName "kube-api-access-tf7l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.657029 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf7l5\" (UniqueName: \"kubernetes.io/projected/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-kube-api-access-tf7l5\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.657066 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:03 crc kubenswrapper[4726]: I1004 03:45:03.657079 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:04 crc kubenswrapper[4726]: I1004 03:45:04.099824 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" event={"ID":"1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf","Type":"ContainerDied","Data":"134abf40b0dd91df9e9b17fba4e0b96591533974747e9ed0ed511d608f1dd221"} Oct 04 03:45:04 crc kubenswrapper[4726]: I1004 03:45:04.099867 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="134abf40b0dd91df9e9b17fba4e0b96591533974747e9ed0ed511d608f1dd221" Oct 04 03:45:04 crc kubenswrapper[4726]: I1004 03:45:04.099891 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn" Oct 04 03:45:19 crc kubenswrapper[4726]: I1004 03:45:19.421383 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.447175 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" containerID="cri-o://16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce" gracePeriod=15 Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.911747 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.963314 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-689bf95868-stn82"] Oct 04 03:45:44 crc kubenswrapper[4726]: E1004 03:45:44.963549 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.963565 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" Oct 04 03:45:44 crc kubenswrapper[4726]: E1004 03:45:44.963587 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" containerName="collect-profiles" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.963597 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" containerName="collect-profiles" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.963699 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" containerName="oauth-openshift" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.963718 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" containerName="collect-profiles" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.964144 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:44 crc kubenswrapper[4726]: I1004 03:45:44.975721 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-689bf95868-stn82"] Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060447 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060561 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060599 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060641 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060680 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060727 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060759 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffdtx\" (UniqueName: \"kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060806 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060873 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060904 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060956 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.060997 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061063 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061136 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data\") pod \"7179a374-528c-4a2f-a648-998a7a5da73e\" (UID: \"7179a374-528c-4a2f-a648-998a7a5da73e\") " Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061303 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-error\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061341 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-dir\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-session\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061428 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061464 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061504 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061543 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-cliconfig\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061576 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-login\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061637 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-router-certs\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-policies\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061667 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061705 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061736 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.061867 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.062219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-service-ca\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.062325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-serving-cert\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.062410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.062815 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063048 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplcq\" (UniqueName: \"kubernetes.io/projected/65e823a8-b1be-4e82-830b-26d8d4f2be12-kube-api-access-gplcq\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063153 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063175 4726 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063195 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063277 4726 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7179a374-528c-4a2f-a648-998a7a5da73e-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.063299 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.067648 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx" (OuterVolumeSpecName: "kube-api-access-ffdtx") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "kube-api-access-ffdtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.068133 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.069095 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.070433 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.073415 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.073841 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.074300 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.074553 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.078475 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7179a374-528c-4a2f-a648-998a7a5da73e" (UID: "7179a374-528c-4a2f-a648-998a7a5da73e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164170 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-cliconfig\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164265 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-login\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164335 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-router-certs\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-policies\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164415 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-service-ca\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164509 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-serving-cert\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplcq\" (UniqueName: \"kubernetes.io/projected/65e823a8-b1be-4e82-830b-26d8d4f2be12-kube-api-access-gplcq\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164628 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-error\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164658 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-dir\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164701 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-session\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164741 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164781 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164816 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164888 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164913 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164934 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164955 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164974 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.164994 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.165014 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.165034 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffdtx\" (UniqueName: \"kubernetes.io/projected/7179a374-528c-4a2f-a648-998a7a5da73e-kube-api-access-ffdtx\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.165059 4726 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7179a374-528c-4a2f-a648-998a7a5da73e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.165839 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-cliconfig\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.166594 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-service-ca\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.167496 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.167790 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-dir\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.169317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/65e823a8-b1be-4e82-830b-26d8d4f2be12-audit-policies\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.170774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.171001 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.171494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-session\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.172391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-router-certs\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.173368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-error\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.173574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-template-login\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.173659 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.174577 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/65e823a8-b1be-4e82-830b-26d8d4f2be12-v4-0-config-system-serving-cert\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.192920 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplcq\" (UniqueName: \"kubernetes.io/projected/65e823a8-b1be-4e82-830b-26d8d4f2be12-kube-api-access-gplcq\") pod \"oauth-openshift-689bf95868-stn82\" (UID: \"65e823a8-b1be-4e82-830b-26d8d4f2be12\") " pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.293781 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.390416 4726 generic.go:334] "Generic (PLEG): container finished" podID="7179a374-528c-4a2f-a648-998a7a5da73e" containerID="16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce" exitCode=0 Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.390933 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" event={"ID":"7179a374-528c-4a2f-a648-998a7a5da73e","Type":"ContainerDied","Data":"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce"} Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.390978 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" event={"ID":"7179a374-528c-4a2f-a648-998a7a5da73e","Type":"ContainerDied","Data":"44eed5c708e332e9e2770a84e1dca6ef22689e17fcd28d2ea8a7856749016187"} Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.391010 4726 scope.go:117] "RemoveContainer" containerID="16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.391248 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wssnh" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.443503 4726 scope.go:117] "RemoveContainer" containerID="16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce" Oct 04 03:45:45 crc kubenswrapper[4726]: E1004 03:45:45.444070 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce\": container with ID starting with 16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce not found: ID does not exist" containerID="16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.444125 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce"} err="failed to get container status \"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce\": rpc error: code = NotFound desc = could not find container \"16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce\": container with ID starting with 16a4b8675946c2d94ce4511fe90800f208933c08517f07ec459b4b045c02b4ce not found: ID does not exist" Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.462602 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.464856 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wssnh"] Oct 04 03:45:45 crc kubenswrapper[4726]: I1004 03:45:45.559005 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-689bf95868-stn82"] Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.402760 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" event={"ID":"65e823a8-b1be-4e82-830b-26d8d4f2be12","Type":"ContainerStarted","Data":"5a3f80b91d3f30ad14275f79a1436203c7016c79755418df21c067d6ca51fec5"} Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.403245 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.403297 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" event={"ID":"65e823a8-b1be-4e82-830b-26d8d4f2be12","Type":"ContainerStarted","Data":"679e25aa09411235353f9aadca17cc76c1111fe4d69365bd76876b4073deeb92"} Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.518842 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7179a374-528c-4a2f-a648-998a7a5da73e" path="/var/lib/kubelet/pods/7179a374-528c-4a2f-a648-998a7a5da73e/volumes" Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.737152 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" Oct 04 03:45:46 crc kubenswrapper[4726]: I1004 03:45:46.759191 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-689bf95868-stn82" podStartSLOduration=27.759174582 podStartE2EDuration="27.759174582s" podCreationTimestamp="2025-10-04 03:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:45:46.437790294 +0000 UTC m=+320.612413537" watchObservedRunningTime="2025-10-04 03:45:46.759174582 +0000 UTC m=+320.933797795" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.658223 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.659127 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-khfd7" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="registry-server" containerID="cri-o://ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70" gracePeriod=30 Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.678693 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.679074 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swwmz" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="registry-server" containerID="cri-o://32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3" gracePeriod=30 Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.699569 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.699813 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" podUID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" containerName="marketplace-operator" containerID="cri-o://8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327" gracePeriod=30 Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.702224 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.702530 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t2vjb" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="registry-server" containerID="cri-o://047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9" gracePeriod=30 Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.707605 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.708137 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lvtrq" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="registry-server" containerID="cri-o://6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753" gracePeriod=30 Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.713090 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mqgpt"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.713975 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.717271 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mqgpt"] Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.774392 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.774443 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q69bd\" (UniqueName: \"kubernetes.io/projected/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-kube-api-access-q69bd\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.774548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.875523 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.875650 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.875680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q69bd\" (UniqueName: \"kubernetes.io/projected/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-kube-api-access-q69bd\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.877608 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.883387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:22 crc kubenswrapper[4726]: I1004 03:46:22.897827 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q69bd\" (UniqueName: \"kubernetes.io/projected/f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8-kube-api-access-q69bd\") pod \"marketplace-operator-79b997595-mqgpt\" (UID: \"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8\") " pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.095333 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.103881 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.108088 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.127304 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.140402 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.141004 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183015 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content\") pod \"c8435186-bbfc-4907-bf0a-2a6b88613f07\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183061 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities\") pod \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183122 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content\") pod \"747e7096-4e73-42d1-8c28-49c4e66d0b19\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktbbt\" (UniqueName: \"kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt\") pod \"747e7096-4e73-42d1-8c28-49c4e66d0b19\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183192 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") pod \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183217 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc6lw\" (UniqueName: \"kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw\") pod \"97de0531-09f7-4b57-b11c-53576c47ba04\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.183282 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxtvd\" (UniqueName: \"kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd\") pod \"c8435186-bbfc-4907-bf0a-2a6b88613f07\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.184369 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") pod \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.185165 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qq46\" (UniqueName: \"kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46\") pod \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\" (UID: \"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.185692 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fmtr\" (UniqueName: \"kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr\") pod \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.185823 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities\") pod \"c8435186-bbfc-4907-bf0a-2a6b88613f07\" (UID: \"c8435186-bbfc-4907-bf0a-2a6b88613f07\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.185902 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities\") pod \"747e7096-4e73-42d1-8c28-49c4e66d0b19\" (UID: \"747e7096-4e73-42d1-8c28-49c4e66d0b19\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.186181 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities\") pod \"97de0531-09f7-4b57-b11c-53576c47ba04\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.186288 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content\") pod \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\" (UID: \"318cd44c-dd4c-4859-ad8b-3f4f308771d0\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.186365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content\") pod \"97de0531-09f7-4b57-b11c-53576c47ba04\" (UID: \"97de0531-09f7-4b57-b11c-53576c47ba04\") " Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.186215 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" (UID: "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.187171 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities" (OuterVolumeSpecName: "utilities") pod "c8435186-bbfc-4907-bf0a-2a6b88613f07" (UID: "c8435186-bbfc-4907-bf0a-2a6b88613f07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.187240 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities" (OuterVolumeSpecName: "utilities") pod "318cd44c-dd4c-4859-ad8b-3f4f308771d0" (UID: "318cd44c-dd4c-4859-ad8b-3f4f308771d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.187294 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities" (OuterVolumeSpecName: "utilities") pod "97de0531-09f7-4b57-b11c-53576c47ba04" (UID: "97de0531-09f7-4b57-b11c-53576c47ba04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.189607 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.189701 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.189761 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.189844 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.192721 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" (UID: "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.192953 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities" (OuterVolumeSpecName: "utilities") pod "747e7096-4e73-42d1-8c28-49c4e66d0b19" (UID: "747e7096-4e73-42d1-8c28-49c4e66d0b19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.193800 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46" (OuterVolumeSpecName: "kube-api-access-9qq46") pod "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" (UID: "b6cc271a-fe0c-40ca-a28d-68568dcdf2d2"). InnerVolumeSpecName "kube-api-access-9qq46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.196428 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw" (OuterVolumeSpecName: "kube-api-access-nc6lw") pod "97de0531-09f7-4b57-b11c-53576c47ba04" (UID: "97de0531-09f7-4b57-b11c-53576c47ba04"). InnerVolumeSpecName "kube-api-access-nc6lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.196560 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd" (OuterVolumeSpecName: "kube-api-access-vxtvd") pod "c8435186-bbfc-4907-bf0a-2a6b88613f07" (UID: "c8435186-bbfc-4907-bf0a-2a6b88613f07"). InnerVolumeSpecName "kube-api-access-vxtvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.208701 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97de0531-09f7-4b57-b11c-53576c47ba04" (UID: "97de0531-09f7-4b57-b11c-53576c47ba04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.210759 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr" (OuterVolumeSpecName: "kube-api-access-8fmtr") pod "318cd44c-dd4c-4859-ad8b-3f4f308771d0" (UID: "318cd44c-dd4c-4859-ad8b-3f4f308771d0"). InnerVolumeSpecName "kube-api-access-8fmtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.210776 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt" (OuterVolumeSpecName: "kube-api-access-ktbbt") pod "747e7096-4e73-42d1-8c28-49c4e66d0b19" (UID: "747e7096-4e73-42d1-8c28-49c4e66d0b19"). InnerVolumeSpecName "kube-api-access-ktbbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.270644 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "747e7096-4e73-42d1-8c28-49c4e66d0b19" (UID: "747e7096-4e73-42d1-8c28-49c4e66d0b19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.288622 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8435186-bbfc-4907-bf0a-2a6b88613f07" (UID: "c8435186-bbfc-4907-bf0a-2a6b88613f07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291154 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxtvd\" (UniqueName: \"kubernetes.io/projected/c8435186-bbfc-4907-bf0a-2a6b88613f07-kube-api-access-vxtvd\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291182 4726 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291193 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qq46\" (UniqueName: \"kubernetes.io/projected/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2-kube-api-access-9qq46\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291202 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fmtr\" (UniqueName: \"kubernetes.io/projected/318cd44c-dd4c-4859-ad8b-3f4f308771d0-kube-api-access-8fmtr\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291213 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291222 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97de0531-09f7-4b57-b11c-53576c47ba04-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291231 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8435186-bbfc-4907-bf0a-2a6b88613f07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291239 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/747e7096-4e73-42d1-8c28-49c4e66d0b19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291247 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktbbt\" (UniqueName: \"kubernetes.io/projected/747e7096-4e73-42d1-8c28-49c4e66d0b19-kube-api-access-ktbbt\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.291255 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc6lw\" (UniqueName: \"kubernetes.io/projected/97de0531-09f7-4b57-b11c-53576c47ba04-kube-api-access-nc6lw\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.299555 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mqgpt"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.315354 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "318cd44c-dd4c-4859-ad8b-3f4f308771d0" (UID: "318cd44c-dd4c-4859-ad8b-3f4f308771d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.392821 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/318cd44c-dd4c-4859-ad8b-3f4f308771d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.664206 4726 generic.go:334] "Generic (PLEG): container finished" podID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerID="32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3" exitCode=0 Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.664298 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwmz" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.664308 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerDied","Data":"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.664361 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwmz" event={"ID":"747e7096-4e73-42d1-8c28-49c4e66d0b19","Type":"ContainerDied","Data":"e3de63caa8dabdc349f7e96c808ccaf92b88bd948764797c7e9266dfc46546ba"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.664385 4726 scope.go:117] "RemoveContainer" containerID="32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.667761 4726 generic.go:334] "Generic (PLEG): container finished" podID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerID="6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753" exitCode=0 Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.667812 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvtrq" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.667851 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerDied","Data":"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.667892 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvtrq" event={"ID":"318cd44c-dd4c-4859-ad8b-3f4f308771d0","Type":"ContainerDied","Data":"6118a21fae8d6e2b84ddd03dc4d46e1781cc768672387cc35d0817f704205b73"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.670827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" event={"ID":"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8","Type":"ContainerStarted","Data":"5997a74e3c934ffa22b1449d5c49c5c849eeb05a514b0eaae144fa342325c4f0"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.670892 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" event={"ID":"f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8","Type":"ContainerStarted","Data":"39923d0c16c49c6438ce528e9336f33b29ebe9f3c125e14520d26f0f13122870"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.671056 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.672444 4726 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mqgpt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.672570 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" podUID="f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.673134 4726 generic.go:334] "Generic (PLEG): container finished" podID="97de0531-09f7-4b57-b11c-53576c47ba04" containerID="047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9" exitCode=0 Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.673234 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerDied","Data":"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.673252 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t2vjb" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.673277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t2vjb" event={"ID":"97de0531-09f7-4b57-b11c-53576c47ba04","Type":"ContainerDied","Data":"b2cb56f7559b73657b60a193d61dfd19bbd1cee82f7c8d18bb109f435b94cbcc"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.681793 4726 generic.go:334] "Generic (PLEG): container finished" podID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" containerID="8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327" exitCode=0 Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.681932 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" event={"ID":"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2","Type":"ContainerDied","Data":"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.682117 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" event={"ID":"b6cc271a-fe0c-40ca-a28d-68568dcdf2d2","Type":"ContainerDied","Data":"b9e0eee9a823ee55a16e205aba87ae059038fd85f635d825e6948f893abe2179"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.682234 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-t8p4b" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.690594 4726 generic.go:334] "Generic (PLEG): container finished" podID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerID="ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70" exitCode=0 Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.690750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerDied","Data":"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.690844 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfd7" event={"ID":"c8435186-bbfc-4907-bf0a-2a6b88613f07","Type":"ContainerDied","Data":"d457e3d7c62d68d3d2b5a65761d490be51139beea124181183ff7a5dc944d26a"} Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.690747 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfd7" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.696380 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" podStartSLOduration=1.6963623220000001 podStartE2EDuration="1.696362322s" podCreationTimestamp="2025-10-04 03:46:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:46:23.693142183 +0000 UTC m=+357.867765426" watchObservedRunningTime="2025-10-04 03:46:23.696362322 +0000 UTC m=+357.870985535" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.701337 4726 scope.go:117] "RemoveContainer" containerID="e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.741628 4726 scope.go:117] "RemoveContainer" containerID="37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.741953 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.745934 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lvtrq"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.770371 4726 scope.go:117] "RemoveContainer" containerID="32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.771199 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3\": container with ID starting with 32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3 not found: ID does not exist" containerID="32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771249 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3"} err="failed to get container status \"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3\": rpc error: code = NotFound desc = could not find container \"32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3\": container with ID starting with 32777c33cf30a7134b598d9f32a9789322afcd35fe3f0454220a17b72e5b63c3 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771276 4726 scope.go:117] "RemoveContainer" containerID="e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.771537 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65\": container with ID starting with e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65 not found: ID does not exist" containerID="e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771566 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65"} err="failed to get container status \"e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65\": rpc error: code = NotFound desc = could not find container \"e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65\": container with ID starting with e5c2db25336121d5ead8f20faf82e4260dda9f20994a89e91effb28bd7deec65 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771586 4726 scope.go:117] "RemoveContainer" containerID="37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.771803 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa\": container with ID starting with 37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa not found: ID does not exist" containerID="37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771825 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa"} err="failed to get container status \"37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa\": rpc error: code = NotFound desc = could not find container \"37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa\": container with ID starting with 37ea13398b0068a26f45fb4c366feab9f1e6892f5c42a61dafcc8ed0c59781fa not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.771842 4726 scope.go:117] "RemoveContainer" containerID="6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.787221 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.814158 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t2vjb"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.822585 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.824686 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swwmz"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.827507 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.828681 4726 scope.go:117] "RemoveContainer" containerID="58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.830085 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-t8p4b"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.832776 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.835018 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-khfd7"] Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.848092 4726 scope.go:117] "RemoveContainer" containerID="85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.868922 4726 scope.go:117] "RemoveContainer" containerID="6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.869373 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753\": container with ID starting with 6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753 not found: ID does not exist" containerID="6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.869486 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753"} err="failed to get container status \"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753\": rpc error: code = NotFound desc = could not find container \"6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753\": container with ID starting with 6490dfa9a70984a422f56a6cc039ec4d311db36f554cf2a7d323702707bd3753 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.869580 4726 scope.go:117] "RemoveContainer" containerID="58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.870014 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3\": container with ID starting with 58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3 not found: ID does not exist" containerID="58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.870133 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3"} err="failed to get container status \"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3\": rpc error: code = NotFound desc = could not find container \"58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3\": container with ID starting with 58f45ac6115b6f8b9ebe7f42e5e628f5a89eb91e923fb5e237230466e6a28cb3 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.870218 4726 scope.go:117] "RemoveContainer" containerID="85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.870798 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70\": container with ID starting with 85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70 not found: ID does not exist" containerID="85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.870848 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70"} err="failed to get container status \"85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70\": rpc error: code = NotFound desc = could not find container \"85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70\": container with ID starting with 85c6c43ea3436476308253b315edd724459cb8da1928c035d4cc1312c84e1b70 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.870881 4726 scope.go:117] "RemoveContainer" containerID="047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.887425 4726 scope.go:117] "RemoveContainer" containerID="d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.902955 4726 scope.go:117] "RemoveContainer" containerID="6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.922521 4726 scope.go:117] "RemoveContainer" containerID="047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.922980 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9\": container with ID starting with 047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9 not found: ID does not exist" containerID="047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.923036 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9"} err="failed to get container status \"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9\": rpc error: code = NotFound desc = could not find container \"047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9\": container with ID starting with 047bdc8fc61cb841c28d26a497651d03a709aca39b7dedbf78033d0b489eb5d9 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.923080 4726 scope.go:117] "RemoveContainer" containerID="d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.923411 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3\": container with ID starting with d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3 not found: ID does not exist" containerID="d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.923440 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3"} err="failed to get container status \"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3\": rpc error: code = NotFound desc = could not find container \"d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3\": container with ID starting with d3b8c50d414f20ef673bd4c91ecc6da8cf82d16e7c40180f25f062b232b211c3 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.923458 4726 scope.go:117] "RemoveContainer" containerID="6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.925140 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d\": container with ID starting with 6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d not found: ID does not exist" containerID="6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.925169 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d"} err="failed to get container status \"6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d\": rpc error: code = NotFound desc = could not find container \"6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d\": container with ID starting with 6fbd34c48bfff5214775909f6d68a48b1a9ad338bc0d99cf5ad7deed7864c13d not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.925187 4726 scope.go:117] "RemoveContainer" containerID="8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.938544 4726 scope.go:117] "RemoveContainer" containerID="8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.939084 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327\": container with ID starting with 8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327 not found: ID does not exist" containerID="8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.939138 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327"} err="failed to get container status \"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327\": rpc error: code = NotFound desc = could not find container \"8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327\": container with ID starting with 8cd4915a9c51cf6f24380c02ed02439ce4b2a0832c83f6ca97e67c45324e4327 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.939159 4726 scope.go:117] "RemoveContainer" containerID="ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.951534 4726 scope.go:117] "RemoveContainer" containerID="e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.965280 4726 scope.go:117] "RemoveContainer" containerID="3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.977455 4726 scope.go:117] "RemoveContainer" containerID="ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.977889 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70\": container with ID starting with ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70 not found: ID does not exist" containerID="ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.977945 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70"} err="failed to get container status \"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70\": rpc error: code = NotFound desc = could not find container \"ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70\": container with ID starting with ff42be02ebab191b8b8e7b4ec19beeb1f9d12804a2293b6797ae16c411c4cf70 not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.977982 4726 scope.go:117] "RemoveContainer" containerID="e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.978573 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f\": container with ID starting with e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f not found: ID does not exist" containerID="e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.978618 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f"} err="failed to get container status \"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f\": rpc error: code = NotFound desc = could not find container \"e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f\": container with ID starting with e4511707e629d1f139df1d103db64ab3eb8f322bfc2e0c9e35a4d89b67713c4f not found: ID does not exist" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.978646 4726 scope.go:117] "RemoveContainer" containerID="3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed" Oct 04 03:46:23 crc kubenswrapper[4726]: E1004 03:46:23.979402 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed\": container with ID starting with 3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed not found: ID does not exist" containerID="3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed" Oct 04 03:46:23 crc kubenswrapper[4726]: I1004 03:46:23.979440 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed"} err="failed to get container status \"3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed\": rpc error: code = NotFound desc = could not find container \"3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed\": container with ID starting with 3d6f297faf36cda7b20eed9f1b9c461385865128ff4e5e44a3e50e71f88ab4ed not found: ID does not exist" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.510809 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" path="/var/lib/kubelet/pods/318cd44c-dd4c-4859-ad8b-3f4f308771d0/volumes" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.512237 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" path="/var/lib/kubelet/pods/747e7096-4e73-42d1-8c28-49c4e66d0b19/volumes" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.513371 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" path="/var/lib/kubelet/pods/97de0531-09f7-4b57-b11c-53576c47ba04/volumes" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.515516 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" path="/var/lib/kubelet/pods/b6cc271a-fe0c-40ca-a28d-68568dcdf2d2/volumes" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.516598 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" path="/var/lib/kubelet/pods/c8435186-bbfc-4907-bf0a-2a6b88613f07/volumes" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.703273 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mqgpt" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875519 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7skgq"] Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875687 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875697 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875709 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875714 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875723 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" containerName="marketplace-operator" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875730 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" containerName="marketplace-operator" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875737 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875743 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875751 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875757 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875763 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875781 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875789 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875794 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875802 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875817 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875825 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875831 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="extract-content" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875843 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.875848 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.875998 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876063 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="extract-utilities" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.876076 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876083 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: E1004 03:46:24.876118 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876124 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876214 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="97de0531-09f7-4b57-b11c-53576c47ba04" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876223 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6cc271a-fe0c-40ca-a28d-68568dcdf2d2" containerName="marketplace-operator" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876231 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="318cd44c-dd4c-4859-ad8b-3f4f308771d0" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876238 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8435186-bbfc-4907-bf0a-2a6b88613f07" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.876248 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="747e7096-4e73-42d1-8c28-49c4e66d0b19" containerName="registry-server" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.878475 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.881088 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.888854 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7skgq"] Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.911799 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-utilities\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.911891 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-catalog-content\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:24 crc kubenswrapper[4726]: I1004 03:46:24.911951 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbw62\" (UniqueName: \"kubernetes.io/projected/b10c732f-90f0-413e-8efe-8ae89ca2ea39-kube-api-access-tbw62\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.013246 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbw62\" (UniqueName: \"kubernetes.io/projected/b10c732f-90f0-413e-8efe-8ae89ca2ea39-kube-api-access-tbw62\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.013381 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-utilities\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.013462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-catalog-content\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.013956 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-utilities\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.014145 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b10c732f-90f0-413e-8efe-8ae89ca2ea39-catalog-content\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.031737 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbw62\" (UniqueName: \"kubernetes.io/projected/b10c732f-90f0-413e-8efe-8ae89ca2ea39-kube-api-access-tbw62\") pod \"redhat-marketplace-7skgq\" (UID: \"b10c732f-90f0-413e-8efe-8ae89ca2ea39\") " pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.075031 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dnxfd"] Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.075981 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.083771 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.086396 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnxfd"] Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.116774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjhmq\" (UniqueName: \"kubernetes.io/projected/c45c00c6-5a4c-4c73-935b-185cf2759ea7-kube-api-access-fjhmq\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.116891 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-catalog-content\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.116975 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-utilities\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.201604 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.218664 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-catalog-content\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.218757 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-utilities\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.218797 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjhmq\" (UniqueName: \"kubernetes.io/projected/c45c00c6-5a4c-4c73-935b-185cf2759ea7-kube-api-access-fjhmq\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.219679 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-catalog-content\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.220354 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c45c00c6-5a4c-4c73-935b-185cf2759ea7-utilities\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.243195 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjhmq\" (UniqueName: \"kubernetes.io/projected/c45c00c6-5a4c-4c73-935b-185cf2759ea7-kube-api-access-fjhmq\") pod \"redhat-operators-dnxfd\" (UID: \"c45c00c6-5a4c-4c73-935b-185cf2759ea7\") " pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.393912 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7skgq"] Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.419664 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.600752 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnxfd"] Oct 04 03:46:25 crc kubenswrapper[4726]: W1004 03:46:25.629532 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc45c00c6_5a4c_4c73_935b_185cf2759ea7.slice/crio-12216a2b7f199a579a625a855f68cdb03812bbb806b3d502187b2ecd1c201250 WatchSource:0}: Error finding container 12216a2b7f199a579a625a855f68cdb03812bbb806b3d502187b2ecd1c201250: Status 404 returned error can't find the container with id 12216a2b7f199a579a625a855f68cdb03812bbb806b3d502187b2ecd1c201250 Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.710847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxfd" event={"ID":"c45c00c6-5a4c-4c73-935b-185cf2759ea7","Type":"ContainerStarted","Data":"12216a2b7f199a579a625a855f68cdb03812bbb806b3d502187b2ecd1c201250"} Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.712598 4726 generic.go:334] "Generic (PLEG): container finished" podID="b10c732f-90f0-413e-8efe-8ae89ca2ea39" containerID="40c96e1bfd8e9806b790f953e757e7b372224218fd8883e41eddd830d52831e9" exitCode=0 Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.712657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7skgq" event={"ID":"b10c732f-90f0-413e-8efe-8ae89ca2ea39","Type":"ContainerDied","Data":"40c96e1bfd8e9806b790f953e757e7b372224218fd8883e41eddd830d52831e9"} Oct 04 03:46:25 crc kubenswrapper[4726]: I1004 03:46:25.712716 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7skgq" event={"ID":"b10c732f-90f0-413e-8efe-8ae89ca2ea39","Type":"ContainerStarted","Data":"79d909bb4ae75c46315c90a97d87ca31d60f0597e3052076998c20d4c469c586"} Oct 04 03:46:26 crc kubenswrapper[4726]: I1004 03:46:26.719859 4726 generic.go:334] "Generic (PLEG): container finished" podID="c45c00c6-5a4c-4c73-935b-185cf2759ea7" containerID="f3fdbc713a6a49642060a53162e79927909e0684ebb6ceb4b0c8eb9eebf0f731" exitCode=0 Oct 04 03:46:26 crc kubenswrapper[4726]: I1004 03:46:26.720025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxfd" event={"ID":"c45c00c6-5a4c-4c73-935b-185cf2759ea7","Type":"ContainerDied","Data":"f3fdbc713a6a49642060a53162e79927909e0684ebb6ceb4b0c8eb9eebf0f731"} Oct 04 03:46:26 crc kubenswrapper[4726]: I1004 03:46:26.732486 4726 generic.go:334] "Generic (PLEG): container finished" podID="b10c732f-90f0-413e-8efe-8ae89ca2ea39" containerID="4d27734306a31c6ea39957f9a588c57381388fe79765308214c46c0be10542d4" exitCode=0 Oct 04 03:46:26 crc kubenswrapper[4726]: I1004 03:46:26.732535 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7skgq" event={"ID":"b10c732f-90f0-413e-8efe-8ae89ca2ea39","Type":"ContainerDied","Data":"4d27734306a31c6ea39957f9a588c57381388fe79765308214c46c0be10542d4"} Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.284502 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-65mdl"] Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.288390 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65mdl"] Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.288429 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.290370 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.350062 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-catalog-content\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.350236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-utilities\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.350314 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxm2j\" (UniqueName: \"kubernetes.io/projected/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-kube-api-access-fxm2j\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.451289 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxm2j\" (UniqueName: \"kubernetes.io/projected/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-kube-api-access-fxm2j\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.451344 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-catalog-content\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.451387 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-utilities\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.451805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-utilities\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.452255 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-catalog-content\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.472819 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxm2j\" (UniqueName: \"kubernetes.io/projected/9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a-kube-api-access-fxm2j\") pod \"community-operators-65mdl\" (UID: \"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a\") " pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.478155 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ptxbw"] Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.479037 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.487314 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.493471 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ptxbw"] Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.553641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-catalog-content\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.554291 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvh92\" (UniqueName: \"kubernetes.io/projected/afe2d9b4-beff-42d2-b599-441fcb58ddbd-kube-api-access-lvh92\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.554533 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-utilities\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.620850 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.655683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-catalog-content\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.655767 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvh92\" (UniqueName: \"kubernetes.io/projected/afe2d9b4-beff-42d2-b599-441fcb58ddbd-kube-api-access-lvh92\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.655806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-utilities\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.656546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-catalog-content\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.656552 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe2d9b4-beff-42d2-b599-441fcb58ddbd-utilities\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.674786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvh92\" (UniqueName: \"kubernetes.io/projected/afe2d9b4-beff-42d2-b599-441fcb58ddbd-kube-api-access-lvh92\") pod \"certified-operators-ptxbw\" (UID: \"afe2d9b4-beff-42d2-b599-441fcb58ddbd\") " pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.741079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7skgq" event={"ID":"b10c732f-90f0-413e-8efe-8ae89ca2ea39","Type":"ContainerStarted","Data":"b50447b8c3773906181cd3f9561d7e241ec3a829c44bc64f413aafee92193022"} Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.754184 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxfd" event={"ID":"c45c00c6-5a4c-4c73-935b-185cf2759ea7","Type":"ContainerStarted","Data":"ae09de24c4a267809563bd814132a3e96effe972041a3b440b40eee081ff4848"} Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.763820 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7skgq" podStartSLOduration=2.315398846 podStartE2EDuration="3.763803996s" podCreationTimestamp="2025-10-04 03:46:24 +0000 UTC" firstStartedPulling="2025-10-04 03:46:25.714155875 +0000 UTC m=+359.888779088" lastFinishedPulling="2025-10-04 03:46:27.162561025 +0000 UTC m=+361.337184238" observedRunningTime="2025-10-04 03:46:27.761248056 +0000 UTC m=+361.935871269" watchObservedRunningTime="2025-10-04 03:46:27.763803996 +0000 UTC m=+361.938427209" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.810577 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.821794 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65mdl"] Oct 04 03:46:27 crc kubenswrapper[4726]: I1004 03:46:27.996291 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ptxbw"] Oct 04 03:46:28 crc kubenswrapper[4726]: W1004 03:46:28.061475 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafe2d9b4_beff_42d2_b599_441fcb58ddbd.slice/crio-a14022c7514f1dff0f7d759448202c74ed153ac18d9a2b24138875e1f0bd405c WatchSource:0}: Error finding container a14022c7514f1dff0f7d759448202c74ed153ac18d9a2b24138875e1f0bd405c: Status 404 returned error can't find the container with id a14022c7514f1dff0f7d759448202c74ed153ac18d9a2b24138875e1f0bd405c Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.761707 4726 generic.go:334] "Generic (PLEG): container finished" podID="9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a" containerID="6a9b5c2e1ff26b6ed091ebf517eec4d8a69974a34e0c08202589dfae220ad4a7" exitCode=0 Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.761880 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65mdl" event={"ID":"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a","Type":"ContainerDied","Data":"6a9b5c2e1ff26b6ed091ebf517eec4d8a69974a34e0c08202589dfae220ad4a7"} Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.761902 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65mdl" event={"ID":"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a","Type":"ContainerStarted","Data":"b5da2b36305ce809dfa52cc20f131c736f7137ca747ea4d29f1d2462d724c6de"} Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.764231 4726 generic.go:334] "Generic (PLEG): container finished" podID="afe2d9b4-beff-42d2-b599-441fcb58ddbd" containerID="e44fa44d3464137fc0cb83ab4736918e085bc0d0819a54c7903d017e2f10d108" exitCode=0 Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.764264 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptxbw" event={"ID":"afe2d9b4-beff-42d2-b599-441fcb58ddbd","Type":"ContainerDied","Data":"e44fa44d3464137fc0cb83ab4736918e085bc0d0819a54c7903d017e2f10d108"} Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.764278 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptxbw" event={"ID":"afe2d9b4-beff-42d2-b599-441fcb58ddbd","Type":"ContainerStarted","Data":"a14022c7514f1dff0f7d759448202c74ed153ac18d9a2b24138875e1f0bd405c"} Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.769836 4726 generic.go:334] "Generic (PLEG): container finished" podID="c45c00c6-5a4c-4c73-935b-185cf2759ea7" containerID="ae09de24c4a267809563bd814132a3e96effe972041a3b440b40eee081ff4848" exitCode=0 Oct 04 03:46:28 crc kubenswrapper[4726]: I1004 03:46:28.769889 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxfd" event={"ID":"c45c00c6-5a4c-4c73-935b-185cf2759ea7","Type":"ContainerDied","Data":"ae09de24c4a267809563bd814132a3e96effe972041a3b440b40eee081ff4848"} Oct 04 03:46:29 crc kubenswrapper[4726]: I1004 03:46:29.776186 4726 generic.go:334] "Generic (PLEG): container finished" podID="9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a" containerID="46cda86bcb26b6c832f402611efa7bfd2089786a064f8e9d213f42cca0f74bdc" exitCode=0 Oct 04 03:46:29 crc kubenswrapper[4726]: I1004 03:46:29.776238 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65mdl" event={"ID":"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a","Type":"ContainerDied","Data":"46cda86bcb26b6c832f402611efa7bfd2089786a064f8e9d213f42cca0f74bdc"} Oct 04 03:46:29 crc kubenswrapper[4726]: I1004 03:46:29.780277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptxbw" event={"ID":"afe2d9b4-beff-42d2-b599-441fcb58ddbd","Type":"ContainerStarted","Data":"bce8dfce8da5628bc14bc86508fc703bcbb94f2462ed8663238c67fc573c8f3d"} Oct 04 03:46:29 crc kubenswrapper[4726]: I1004 03:46:29.782657 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxfd" event={"ID":"c45c00c6-5a4c-4c73-935b-185cf2759ea7","Type":"ContainerStarted","Data":"d5915330a3264896ad3c708217596048e108e2b97470198fe04e05914abc3b14"} Oct 04 03:46:29 crc kubenswrapper[4726]: I1004 03:46:29.810070 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dnxfd" podStartSLOduration=2.205800517 podStartE2EDuration="4.810051304s" podCreationTimestamp="2025-10-04 03:46:25 +0000 UTC" firstStartedPulling="2025-10-04 03:46:26.721998212 +0000 UTC m=+360.896621455" lastFinishedPulling="2025-10-04 03:46:29.326249039 +0000 UTC m=+363.500872242" observedRunningTime="2025-10-04 03:46:29.806530187 +0000 UTC m=+363.981153400" watchObservedRunningTime="2025-10-04 03:46:29.810051304 +0000 UTC m=+363.984674517" Oct 04 03:46:30 crc kubenswrapper[4726]: I1004 03:46:30.789982 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65mdl" event={"ID":"9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a","Type":"ContainerStarted","Data":"063238b16f817fe3fcb4ef6b39d383b38eec5ec9471eb26730b07122ced756c8"} Oct 04 03:46:30 crc kubenswrapper[4726]: I1004 03:46:30.792371 4726 generic.go:334] "Generic (PLEG): container finished" podID="afe2d9b4-beff-42d2-b599-441fcb58ddbd" containerID="bce8dfce8da5628bc14bc86508fc703bcbb94f2462ed8663238c67fc573c8f3d" exitCode=0 Oct 04 03:46:30 crc kubenswrapper[4726]: I1004 03:46:30.792409 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptxbw" event={"ID":"afe2d9b4-beff-42d2-b599-441fcb58ddbd","Type":"ContainerDied","Data":"bce8dfce8da5628bc14bc86508fc703bcbb94f2462ed8663238c67fc573c8f3d"} Oct 04 03:46:30 crc kubenswrapper[4726]: I1004 03:46:30.813514 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-65mdl" podStartSLOduration=2.388740701 podStartE2EDuration="3.81349255s" podCreationTimestamp="2025-10-04 03:46:27 +0000 UTC" firstStartedPulling="2025-10-04 03:46:28.763805467 +0000 UTC m=+362.938428670" lastFinishedPulling="2025-10-04 03:46:30.188557266 +0000 UTC m=+364.363180519" observedRunningTime="2025-10-04 03:46:30.811454144 +0000 UTC m=+364.986077347" watchObservedRunningTime="2025-10-04 03:46:30.81349255 +0000 UTC m=+364.988115773" Oct 04 03:46:32 crc kubenswrapper[4726]: I1004 03:46:32.803865 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ptxbw" event={"ID":"afe2d9b4-beff-42d2-b599-441fcb58ddbd","Type":"ContainerStarted","Data":"73d1ba17da15952ceccbe3229c1d8f59d3d806d553436991a05cf3a53815c8c7"} Oct 04 03:46:32 crc kubenswrapper[4726]: I1004 03:46:32.825591 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ptxbw" podStartSLOduration=3.3140836240000002 podStartE2EDuration="5.825576566s" podCreationTimestamp="2025-10-04 03:46:27 +0000 UTC" firstStartedPulling="2025-10-04 03:46:28.766058029 +0000 UTC m=+362.940681242" lastFinishedPulling="2025-10-04 03:46:31.277550971 +0000 UTC m=+365.452174184" observedRunningTime="2025-10-04 03:46:32.824006613 +0000 UTC m=+366.998629826" watchObservedRunningTime="2025-10-04 03:46:32.825576566 +0000 UTC m=+367.000199779" Oct 04 03:46:34 crc kubenswrapper[4726]: I1004 03:46:34.188496 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:46:34 crc kubenswrapper[4726]: I1004 03:46:34.188913 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.202065 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.202219 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.259434 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.419865 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.419935 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.464189 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.859422 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7skgq" Oct 04 03:46:35 crc kubenswrapper[4726]: I1004 03:46:35.877020 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dnxfd" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.621596 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.621944 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.661929 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.812269 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.812358 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.863153 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.880802 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-65mdl" Oct 04 03:46:37 crc kubenswrapper[4726]: I1004 03:46:37.902358 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ptxbw" Oct 04 03:47:04 crc kubenswrapper[4726]: I1004 03:47:04.189162 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:47:04 crc kubenswrapper[4726]: I1004 03:47:04.189909 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:47:34 crc kubenswrapper[4726]: I1004 03:47:34.188402 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:47:34 crc kubenswrapper[4726]: I1004 03:47:34.191243 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:47:34 crc kubenswrapper[4726]: I1004 03:47:34.191321 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:47:34 crc kubenswrapper[4726]: I1004 03:47:34.192426 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:47:34 crc kubenswrapper[4726]: I1004 03:47:34.192568 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e" gracePeriod=600 Oct 04 03:47:35 crc kubenswrapper[4726]: I1004 03:47:35.206724 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e" exitCode=0 Oct 04 03:47:35 crc kubenswrapper[4726]: I1004 03:47:35.206774 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e"} Oct 04 03:47:35 crc kubenswrapper[4726]: I1004 03:47:35.207405 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41"} Oct 04 03:47:35 crc kubenswrapper[4726]: I1004 03:47:35.207436 4726 scope.go:117] "RemoveContainer" containerID="b1f1d828568eea1a3dfe35a673e699bae688a2defb01da1361674c8581ac4ecc" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.233782 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dfxjn"] Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.235476 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.252590 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dfxjn"] Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.403926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-certificates\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.403971 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0ac3404f-7277-47c2-9a41-4be8adbb2ace-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.403993 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp8tn\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-kube-api-access-pp8tn\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.404217 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-trusted-ca\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.404336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-bound-sa-token\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.404428 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.404471 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0ac3404f-7277-47c2-9a41-4be8adbb2ace-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.404513 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-tls\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.438684 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505293 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-trusted-ca\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505352 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-bound-sa-token\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505398 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0ac3404f-7277-47c2-9a41-4be8adbb2ace-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-tls\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-certificates\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505495 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0ac3404f-7277-47c2-9a41-4be8adbb2ace-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.505518 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp8tn\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-kube-api-access-pp8tn\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.507266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-certificates\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.507167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/0ac3404f-7277-47c2-9a41-4be8adbb2ace-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.509292 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0ac3404f-7277-47c2-9a41-4be8adbb2ace-trusted-ca\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.514419 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-registry-tls\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.515899 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/0ac3404f-7277-47c2-9a41-4be8adbb2ace-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.523315 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp8tn\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-kube-api-access-pp8tn\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.524503 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0ac3404f-7277-47c2-9a41-4be8adbb2ace-bound-sa-token\") pod \"image-registry-66df7c8f76-dfxjn\" (UID: \"0ac3404f-7277-47c2-9a41-4be8adbb2ace\") " pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.558333 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:04 crc kubenswrapper[4726]: I1004 03:48:04.777934 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dfxjn"] Oct 04 03:48:05 crc kubenswrapper[4726]: I1004 03:48:05.406854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" event={"ID":"0ac3404f-7277-47c2-9a41-4be8adbb2ace","Type":"ContainerStarted","Data":"738c16cd2f9e3f6e7c93f4fd5632e0002ed6db620f25930ddb905291a503414b"} Oct 04 03:48:05 crc kubenswrapper[4726]: I1004 03:48:05.407186 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:05 crc kubenswrapper[4726]: I1004 03:48:05.407203 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" event={"ID":"0ac3404f-7277-47c2-9a41-4be8adbb2ace","Type":"ContainerStarted","Data":"6e6517074edacf4750ed13b87c9a309337c84c2612850ffccb7978df844ca01f"} Oct 04 03:48:05 crc kubenswrapper[4726]: I1004 03:48:05.436758 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" podStartSLOduration=1.436737456 podStartE2EDuration="1.436737456s" podCreationTimestamp="2025-10-04 03:48:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:48:05.433539573 +0000 UTC m=+459.608162806" watchObservedRunningTime="2025-10-04 03:48:05.436737456 +0000 UTC m=+459.611360669" Oct 04 03:48:24 crc kubenswrapper[4726]: I1004 03:48:24.567346 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dfxjn" Oct 04 03:48:24 crc kubenswrapper[4726]: I1004 03:48:24.630859 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:48:49 crc kubenswrapper[4726]: I1004 03:48:49.680079 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" podUID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" containerName="registry" containerID="cri-o://8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60" gracePeriod=30 Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.110027 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.185948 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186041 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186099 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsq4j\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186253 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186382 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186640 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186698 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.186965 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\" (UID: \"a88cfddf-a8f8-46a9-b581-7993f7d85ef4\") " Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.187444 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.188127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.194592 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.195976 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.197001 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.202754 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.203006 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j" (OuterVolumeSpecName: "kube-api-access-xsq4j") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "kube-api-access-xsq4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.208783 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a88cfddf-a8f8-46a9-b581-7993f7d85ef4" (UID: "a88cfddf-a8f8-46a9-b581-7993f7d85ef4"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288691 4726 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288750 4726 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288771 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsq4j\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-kube-api-access-xsq4j\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288790 4726 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288810 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288830 4726 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.288848 4726 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a88cfddf-a8f8-46a9-b581-7993f7d85ef4-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.697194 4726 generic.go:334] "Generic (PLEG): container finished" podID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" containerID="8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60" exitCode=0 Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.697438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" event={"ID":"a88cfddf-a8f8-46a9-b581-7993f7d85ef4","Type":"ContainerDied","Data":"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60"} Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.697609 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" event={"ID":"a88cfddf-a8f8-46a9-b581-7993f7d85ef4","Type":"ContainerDied","Data":"248fca86822452501376586253a7fe9b13f5e664426ddeb5eb8e416bcea8c763"} Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.697648 4726 scope.go:117] "RemoveContainer" containerID="8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.697469 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-gq29w" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.735720 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.745546 4726 scope.go:117] "RemoveContainer" containerID="8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60" Oct 04 03:48:50 crc kubenswrapper[4726]: E1004 03:48:50.746180 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60\": container with ID starting with 8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60 not found: ID does not exist" containerID="8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.746256 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60"} err="failed to get container status \"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60\": rpc error: code = NotFound desc = could not find container \"8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60\": container with ID starting with 8e5d366523d3896ce413034fab70cac7db646d46b4572c2386a027c60f687a60 not found: ID does not exist" Oct 04 03:48:50 crc kubenswrapper[4726]: I1004 03:48:50.746832 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-gq29w"] Oct 04 03:48:52 crc kubenswrapper[4726]: I1004 03:48:52.513018 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" path="/var/lib/kubelet/pods/a88cfddf-a8f8-46a9-b581-7993f7d85ef4/volumes" Oct 04 03:49:34 crc kubenswrapper[4726]: I1004 03:49:34.188414 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:49:34 crc kubenswrapper[4726]: I1004 03:49:34.189086 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:50:04 crc kubenswrapper[4726]: I1004 03:50:04.188566 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:50:04 crc kubenswrapper[4726]: I1004 03:50:04.189251 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.190100 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.190867 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.190925 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.191871 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.191973 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41" gracePeriod=600 Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.394002 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41" exitCode=0 Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.394061 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41"} Oct 04 03:50:34 crc kubenswrapper[4726]: I1004 03:50:34.394107 4726 scope.go:117] "RemoveContainer" containerID="f43cd634986c37467e553bac686e0cf57cf88df3b9dbc57d4c4aaad977ad6b7e" Oct 04 03:50:35 crc kubenswrapper[4726]: I1004 03:50:35.404255 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db"} Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.874955 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hkk79"] Oct 04 03:52:28 crc kubenswrapper[4726]: E1004 03:52:28.876708 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" containerName="registry" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.876801 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" containerName="registry" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.876997 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88cfddf-a8f8-46a9-b581-7993f7d85ef4" containerName="registry" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.877586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.880437 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.880478 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-6xj49" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.880646 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.890538 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vbkwl"] Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.891597 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vbkwl" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.894247 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8jth9" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.902589 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vbkwl"] Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.921613 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hkk79"] Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.934120 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkt2j"] Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.934737 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.936415 4726 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-5w2mn" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.964705 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brnqk\" (UniqueName: \"kubernetes.io/projected/5e7b44b7-601a-4d34-a2f5-13662f3ef3cd-kube-api-access-brnqk\") pod \"cert-manager-webhook-5655c58dd6-nkt2j\" (UID: \"5e7b44b7-601a-4d34-a2f5-13662f3ef3cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.964751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56j8w\" (UniqueName: \"kubernetes.io/projected/3628e5db-697e-495c-815f-c08b5ccd03da-kube-api-access-56j8w\") pod \"cert-manager-cainjector-7f985d654d-hkk79\" (UID: \"3628e5db-697e-495c-815f-c08b5ccd03da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.964774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9bvw\" (UniqueName: \"kubernetes.io/projected/53efe0a4-7f9a-4541-9677-12e0dcec292e-kube-api-access-r9bvw\") pod \"cert-manager-5b446d88c5-vbkwl\" (UID: \"53efe0a4-7f9a-4541-9677-12e0dcec292e\") " pod="cert-manager/cert-manager-5b446d88c5-vbkwl" Oct 04 03:52:28 crc kubenswrapper[4726]: I1004 03:52:28.975688 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkt2j"] Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.065447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brnqk\" (UniqueName: \"kubernetes.io/projected/5e7b44b7-601a-4d34-a2f5-13662f3ef3cd-kube-api-access-brnqk\") pod \"cert-manager-webhook-5655c58dd6-nkt2j\" (UID: \"5e7b44b7-601a-4d34-a2f5-13662f3ef3cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.065504 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56j8w\" (UniqueName: \"kubernetes.io/projected/3628e5db-697e-495c-815f-c08b5ccd03da-kube-api-access-56j8w\") pod \"cert-manager-cainjector-7f985d654d-hkk79\" (UID: \"3628e5db-697e-495c-815f-c08b5ccd03da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.065533 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9bvw\" (UniqueName: \"kubernetes.io/projected/53efe0a4-7f9a-4541-9677-12e0dcec292e-kube-api-access-r9bvw\") pod \"cert-manager-5b446d88c5-vbkwl\" (UID: \"53efe0a4-7f9a-4541-9677-12e0dcec292e\") " pod="cert-manager/cert-manager-5b446d88c5-vbkwl" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.083199 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brnqk\" (UniqueName: \"kubernetes.io/projected/5e7b44b7-601a-4d34-a2f5-13662f3ef3cd-kube-api-access-brnqk\") pod \"cert-manager-webhook-5655c58dd6-nkt2j\" (UID: \"5e7b44b7-601a-4d34-a2f5-13662f3ef3cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.083252 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56j8w\" (UniqueName: \"kubernetes.io/projected/3628e5db-697e-495c-815f-c08b5ccd03da-kube-api-access-56j8w\") pod \"cert-manager-cainjector-7f985d654d-hkk79\" (UID: \"3628e5db-697e-495c-815f-c08b5ccd03da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.083306 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9bvw\" (UniqueName: \"kubernetes.io/projected/53efe0a4-7f9a-4541-9677-12e0dcec292e-kube-api-access-r9bvw\") pod \"cert-manager-5b446d88c5-vbkwl\" (UID: \"53efe0a4-7f9a-4541-9677-12e0dcec292e\") " pod="cert-manager/cert-manager-5b446d88c5-vbkwl" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.196390 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.207660 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vbkwl" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.253406 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.400977 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hkk79"] Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.413771 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.442263 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vbkwl"] Oct 04 03:52:29 crc kubenswrapper[4726]: W1004 03:52:29.447258 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53efe0a4_7f9a_4541_9677_12e0dcec292e.slice/crio-3a3ccfdee0c99f1f78b43de7b1d3b2bd9cb71780f744979dde01d3a13b789435 WatchSource:0}: Error finding container 3a3ccfdee0c99f1f78b43de7b1d3b2bd9cb71780f744979dde01d3a13b789435: Status 404 returned error can't find the container with id 3a3ccfdee0c99f1f78b43de7b1d3b2bd9cb71780f744979dde01d3a13b789435 Oct 04 03:52:29 crc kubenswrapper[4726]: I1004 03:52:29.483693 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-nkt2j"] Oct 04 03:52:29 crc kubenswrapper[4726]: W1004 03:52:29.488758 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e7b44b7_601a_4d34_a2f5_13662f3ef3cd.slice/crio-923cb3b9432a28ee6ce921c3ea07bc4b5ba3b371075b57372e27c7e68c625382 WatchSource:0}: Error finding container 923cb3b9432a28ee6ce921c3ea07bc4b5ba3b371075b57372e27c7e68c625382: Status 404 returned error can't find the container with id 923cb3b9432a28ee6ce921c3ea07bc4b5ba3b371075b57372e27c7e68c625382 Oct 04 03:52:30 crc kubenswrapper[4726]: I1004 03:52:30.240609 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vbkwl" event={"ID":"53efe0a4-7f9a-4541-9677-12e0dcec292e","Type":"ContainerStarted","Data":"3a3ccfdee0c99f1f78b43de7b1d3b2bd9cb71780f744979dde01d3a13b789435"} Oct 04 03:52:30 crc kubenswrapper[4726]: I1004 03:52:30.241923 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" event={"ID":"5e7b44b7-601a-4d34-a2f5-13662f3ef3cd","Type":"ContainerStarted","Data":"923cb3b9432a28ee6ce921c3ea07bc4b5ba3b371075b57372e27c7e68c625382"} Oct 04 03:52:30 crc kubenswrapper[4726]: I1004 03:52:30.243119 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" event={"ID":"3628e5db-697e-495c-815f-c08b5ccd03da","Type":"ContainerStarted","Data":"30ba9d90def3c160c2ff5ef9da1990def5923d1c97be99b00aaab62f3e595875"} Oct 04 03:52:34 crc kubenswrapper[4726]: I1004 03:52:34.188739 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:52:34 crc kubenswrapper[4726]: I1004 03:52:34.189200 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.233309 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-snjd6"] Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234638 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-controller" containerID="cri-o://2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234701 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="sbdb" containerID="cri-o://d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234762 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="northd" containerID="cri-o://8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234861 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" containerID="cri-o://3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234896 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="nbdb" containerID="cri-o://6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234926 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-node" containerID="cri-o://7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.234896 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f" gracePeriod=30 Oct 04 03:52:39 crc kubenswrapper[4726]: I1004 03:52:39.295939 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" containerID="cri-o://66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" gracePeriod=30 Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.323508 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.325185 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/1.log" Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.329058 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/0.log" Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.330343 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6" exitCode=143 Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.330412 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6"} Oct 04 03:52:40 crc kubenswrapper[4726]: I1004 03:52:40.330473 4726 scope.go:117] "RemoveContainer" containerID="c0923bcbbb15b1174e3976051a2b166b27af631ec91f2a884554eecf1f4ee904" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.338672 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.339815 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/1.log" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344373 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-controller/0.log" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344729 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344770 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344785 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344834 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344848 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344867 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5" exitCode=0 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344881 4726 generic.go:334] "Generic (PLEG): container finished" podID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerID="2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb" exitCode=143 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344947 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.344988 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.345009 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.345028 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.345048 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.345065 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.345084 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.347056 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/2.log" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.347579 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/1.log" Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.347630 4726 generic.go:334] "Generic (PLEG): container finished" podID="7fa9344a-8c9c-4bc5-b91e-b6c1400d4025" containerID="5db07f4304cf2eab69f21bcceb0014c80a2a876d97a3ce476e76b3e8dc8880d6" exitCode=2 Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.347658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerDied","Data":"5db07f4304cf2eab69f21bcceb0014c80a2a876d97a3ce476e76b3e8dc8880d6"} Oct 04 03:52:41 crc kubenswrapper[4726]: I1004 03:52:41.348225 4726 scope.go:117] "RemoveContainer" containerID="5db07f4304cf2eab69f21bcceb0014c80a2a876d97a3ce476e76b3e8dc8880d6" Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.743219 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e is running failed: container process not found" containerID="d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.743402 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3 is running failed: container process not found" containerID="6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.744377 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e is running failed: container process not found" containerID="d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.744425 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3 is running failed: container process not found" containerID="6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.745071 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3 is running failed: container process not found" containerID="6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.745165 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="nbdb" Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.745215 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e is running failed: container process not found" containerID="d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Oct 04 03:52:48 crc kubenswrapper[4726]: E1004 03:52:48.745306 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="sbdb" Oct 04 03:52:50 crc kubenswrapper[4726]: E1004 03:52:50.246551 4726 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.745s" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.113046 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188 is running failed: container process not found" containerID="66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.113716 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188 is running failed: container process not found" containerID="66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.114511 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188 is running failed: container process not found" containerID="66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.114591 4726 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.511035 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovnkube-controller/3.log" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.512257 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/1.log" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.515781 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-controller/0.log" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.516696 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.581920 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.581969 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.581991 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582025 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582040 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582074 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqlqr\" (UniqueName: \"kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582090 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582179 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582200 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582217 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582239 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582262 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582295 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582310 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582332 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582353 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.582374 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert\") pod \"a8d6c450-3310-4058-90dc-0a39ab1934c4\" (UID: \"a8d6c450-3310-4058-90dc-0a39ab1934c4\") " Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583193 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583306 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash" (OuterVolumeSpecName: "host-slash") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583395 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583748 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log" (OuterVolumeSpecName: "node-log") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583816 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.583927 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584229 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584316 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584339 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket" (OuterVolumeSpecName: "log-socket") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584362 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.584956 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.585245 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.585389 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.587542 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2ssfl"] Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.587998 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588023 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="northd" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588257 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="northd" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588338 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588401 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588473 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588533 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588591 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588652 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588719 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588780 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.588847 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588947 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589121 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="nbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589199 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="nbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589267 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589367 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589452 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589522 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589676 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-node" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589755 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-node" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588157 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.588327 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589824 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="sbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589946 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="sbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589964 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kubecfg-setup" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589974 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kubecfg-setup" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.589985 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.589993 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592647 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592670 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="nbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592680 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="sbdb" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592690 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592696 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-node" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592702 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592710 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592717 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592725 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="northd" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592733 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-acl-logging" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592740 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovn-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592750 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592763 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: E1004 03:52:51.592879 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.592890 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" containerName="ovnkube-controller" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.594721 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.608922 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.609449 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr" (OuterVolumeSpecName: "kube-api-access-xqlqr") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "kube-api-access-xqlqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.610068 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a8d6c450-3310-4058-90dc-0a39ab1934c4" (UID: "a8d6c450-3310-4058-90dc-0a39ab1934c4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.682884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-script-lib\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.682978 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fad4d495-d9e2-4428-a929-c9816972ac31-ovn-node-metrics-cert\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683008 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683036 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683058 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-node-log\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683074 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-etc-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683093 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-var-lib-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683120 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-netns\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683137 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-systemd-units\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683258 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683320 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-kubelet\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-bin\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683371 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-slash\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-systemd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-log-socket\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683429 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-netd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683449 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-ovn\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683464 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-config\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683482 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-env-overrides\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683500 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r74d\" (UniqueName: \"kubernetes.io/projected/fad4d495-d9e2-4428-a929-c9816972ac31-kube-api-access-7r74d\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683543 4726 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683555 4726 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683564 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683573 4726 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683582 4726 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683589 4726 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683597 4726 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683606 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683613 4726 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683622 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683630 4726 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683638 4726 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683645 4726 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683654 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683664 4726 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683673 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqlqr\" (UniqueName: \"kubernetes.io/projected/a8d6c450-3310-4058-90dc-0a39ab1934c4-kube-api-access-xqlqr\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683682 4726 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683690 4726 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683698 4726 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a8d6c450-3310-4058-90dc-0a39ab1934c4-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.683707 4726 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a8d6c450-3310-4058-90dc-0a39ab1934c4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fad4d495-d9e2-4428-a929-c9816972ac31-ovn-node-metrics-cert\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784181 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-node-log\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784199 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-etc-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-var-lib-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784237 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-netns\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784252 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-systemd-units\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784287 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-kubelet\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784301 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-bin\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784306 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-etc-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784333 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-node-log\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784354 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-slash\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784358 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784386 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-var-lib-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784395 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-run-netns\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-slash\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784402 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-kubelet\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784306 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784394 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-bin\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-systemd-units\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-openvswitch\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784461 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-systemd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784493 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-systemd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784517 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-log-socket\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784572 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-netd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784600 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-log-socket\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784605 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-ovn\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784632 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-run-ovn\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784670 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fad4d495-d9e2-4428-a929-c9816972ac31-host-cni-netd\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784681 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-config\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784732 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-env-overrides\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784768 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r74d\" (UniqueName: \"kubernetes.io/projected/fad4d495-d9e2-4428-a929-c9816972ac31-kube-api-access-7r74d\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.784800 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-script-lib\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.785332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-env-overrides\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.785579 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-script-lib\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.785665 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fad4d495-d9e2-4428-a929-c9816972ac31-ovnkube-config\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.789757 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fad4d495-d9e2-4428-a929-c9816972ac31-ovn-node-metrics-cert\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.808911 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r74d\" (UniqueName: \"kubernetes.io/projected/fad4d495-d9e2-4428-a929-c9816972ac31-kube-api-access-7r74d\") pod \"ovnkube-node-2ssfl\" (UID: \"fad4d495-d9e2-4428-a929-c9816972ac31\") " pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:51 crc kubenswrapper[4726]: I1004 03:52:51.952589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.166328 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.436071 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/2.log" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.436477 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/1.log" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.436541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5c96v" event={"ID":"7fa9344a-8c9c-4bc5-b91e-b6c1400d4025","Type":"ContainerStarted","Data":"3386379c620f232d5e0c382ed8395f9fc475acb054c06c738fc1f67f703c6c5d"} Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.439008 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/1.log" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.441706 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-controller/0.log" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.442281 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" event={"ID":"a8d6c450-3310-4058-90dc-0a39ab1934c4","Type":"ContainerDied","Data":"cade49fd4bdcc0cebe8a3471eaa37abe150a63fcdb9330da59719a47e6209fed"} Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.442368 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-snjd6" Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.515316 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-snjd6"] Oct 04 03:52:52 crc kubenswrapper[4726]: I1004 03:52:52.518606 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-snjd6"] Oct 04 03:52:53 crc kubenswrapper[4726]: I1004 03:52:53.807762 4726 scope.go:117] "RemoveContainer" containerID="f76c1859cdd6cd72b6a50c6f1e576c403668ed1695308c242220415b7629fee4" Oct 04 03:52:54 crc kubenswrapper[4726]: I1004 03:52:54.460571 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-acl-logging/1.log" Oct 04 03:52:54 crc kubenswrapper[4726]: I1004 03:52:54.464677 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-controller/0.log" Oct 04 03:52:54 crc kubenswrapper[4726]: I1004 03:52:54.516543 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8d6c450-3310-4058-90dc-0a39ab1934c4" path="/var/lib/kubelet/pods/a8d6c450-3310-4058-90dc-0a39ab1934c4/volumes" Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.370970 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.372504 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" containerID="cri-o://d90f0bb0450edf32ee4ebcc8370e81d6b84509f0f507f4bec08d62d2b3c8eff7" gracePeriod=30 Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.472090 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.472365 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" containerID="cri-o://5e78fa72a941428ddbc806f3eebc4904263cb430ef50f4cb7a4cbefeb19b9562" gracePeriod=30 Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.575310 4726 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pbtx9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 03:52:56 crc kubenswrapper[4726]: I1004 03:52:56.575361 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.440087 4726 scope.go:117] "RemoveContainer" containerID="66130998b00378abdc84d322de9937248571161e075f06389ac285a899368188" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.489196 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5c96v_7fa9344a-8c9c-4bc5-b91e-b6c1400d4025/kube-multus/2.log" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.492537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"eec70f68fe5ceaf6061a0b9981763ec4065a36630e0b5b683483a161b244e7f2"} Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.497984 4726 generic.go:334] "Generic (PLEG): container finished" podID="c2ca287e-027c-44b2-8068-b84437de7e71" containerID="5e78fa72a941428ddbc806f3eebc4904263cb430ef50f4cb7a4cbefeb19b9562" exitCode=0 Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.498063 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" event={"ID":"c2ca287e-027c-44b2-8068-b84437de7e71","Type":"ContainerDied","Data":"5e78fa72a941428ddbc806f3eebc4904263cb430ef50f4cb7a4cbefeb19b9562"} Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.500256 4726 generic.go:334] "Generic (PLEG): container finished" podID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerID="d90f0bb0450edf32ee4ebcc8370e81d6b84509f0f507f4bec08d62d2b3c8eff7" exitCode=0 Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.500287 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" event={"ID":"0854df0d-1ca6-41a4-ba86-98fdec338fc1","Type":"ContainerDied","Data":"d90f0bb0450edf32ee4ebcc8370e81d6b84509f0f507f4bec08d62d2b3c8eff7"} Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.861859 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.891628 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-689c849788-ml4dx"] Oct 04 03:52:57 crc kubenswrapper[4726]: E1004 03:52:57.891830 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.891840 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.891924 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" containerName="controller-manager" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.892357 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970617 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config\") pod \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970689 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca\") pod \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970718 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glqrt\" (UniqueName: \"kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt\") pod \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970754 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert\") pod \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970784 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles\") pod \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\" (UID: \"0854df0d-1ca6-41a4-ba86-98fdec338fc1\") " Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970934 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-serving-cert\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.970973 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zz8\" (UniqueName: \"kubernetes.io/projected/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-kube-api-access-97zz8\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-proxy-ca-bundles\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-client-ca\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971059 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-config\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971604 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0854df0d-1ca6-41a4-ba86-98fdec338fc1" (UID: "0854df0d-1ca6-41a4-ba86-98fdec338fc1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971714 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca" (OuterVolumeSpecName: "client-ca") pod "0854df0d-1ca6-41a4-ba86-98fdec338fc1" (UID: "0854df0d-1ca6-41a4-ba86-98fdec338fc1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.971809 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config" (OuterVolumeSpecName: "config") pod "0854df0d-1ca6-41a4-ba86-98fdec338fc1" (UID: "0854df0d-1ca6-41a4-ba86-98fdec338fc1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.977609 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt" (OuterVolumeSpecName: "kube-api-access-glqrt") pod "0854df0d-1ca6-41a4-ba86-98fdec338fc1" (UID: "0854df0d-1ca6-41a4-ba86-98fdec338fc1"). InnerVolumeSpecName "kube-api-access-glqrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:52:57 crc kubenswrapper[4726]: I1004 03:52:57.979989 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0854df0d-1ca6-41a4-ba86-98fdec338fc1" (UID: "0854df0d-1ca6-41a4-ba86-98fdec338fc1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.071767 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-proxy-ca-bundles\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.071862 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-client-ca\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.071924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-config\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.071995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-serving-cert\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072030 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97zz8\" (UniqueName: \"kubernetes.io/projected/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-kube-api-access-97zz8\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072137 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072159 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072178 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glqrt\" (UniqueName: \"kubernetes.io/projected/0854df0d-1ca6-41a4-ba86-98fdec338fc1-kube-api-access-glqrt\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072197 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0854df0d-1ca6-41a4-ba86-98fdec338fc1-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072215 4726 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0854df0d-1ca6-41a4-ba86-98fdec338fc1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.072871 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-proxy-ca-bundles\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.074738 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-config\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.074748 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-client-ca\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.078963 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-serving-cert\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.091391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97zz8\" (UniqueName: \"kubernetes.io/projected/d1beb981-cc80-4a6f-9ac3-cc837cc7b897-kube-api-access-97zz8\") pod \"controller-manager-689c849788-ml4dx\" (UID: \"d1beb981-cc80-4a6f-9ac3-cc837cc7b897\") " pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: E1004 03:52:58.178514 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="quay.io/jetstack/cert-manager-webhook:v1.14.4" Oct 04 03:52:58 crc kubenswrapper[4726]: E1004 03:52:58.178764 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-webhook,Image:quay.io/jetstack/cert-manager-webhook:v1.14.4,Command:[],Args:[--v=2 --secure-port=10250 --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) --dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-dns-names=cert-manager-webhook --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE) --dynamic-serving-dns-names=cert-manager-webhook.$(POD_NAMESPACE).svc],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:10250,Protocol:TCP,HostIP:,},ContainerPort{Name:healthcheck,HostPort:0,ContainerPort:6080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brnqk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 6080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:60,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 6080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000680000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-webhook-5655c58dd6-nkt2j_cert-manager(5e7b44b7-601a-4d34-a2f5-13662f3ef3cd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 03:52:58 crc kubenswrapper[4726]: E1004 03:52:58.180073 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" podUID="5e7b44b7-601a-4d34-a2f5-13662f3ef3cd" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.205435 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.212055 4726 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ttws2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.212185 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.506587 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.508936 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pbtx9" event={"ID":"0854df0d-1ca6-41a4-ba86-98fdec338fc1","Type":"ContainerDied","Data":"d3e0713ba5255ff56017c441115a57eb04ff28ef8b611a8ca8c7aa4bd7ee0ea0"} Oct 04 03:52:58 crc kubenswrapper[4726]: E1004 03:52:58.540465 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/jetstack/cert-manager-webhook:v1.14.4\\\"\"" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" podUID="5e7b44b7-601a-4d34-a2f5-13662f3ef3cd" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.551762 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.554638 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pbtx9"] Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.562614 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.679433 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6g7m\" (UniqueName: \"kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m\") pod \"c2ca287e-027c-44b2-8068-b84437de7e71\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.679554 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert\") pod \"c2ca287e-027c-44b2-8068-b84437de7e71\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.679609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config\") pod \"c2ca287e-027c-44b2-8068-b84437de7e71\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.679635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca\") pod \"c2ca287e-027c-44b2-8068-b84437de7e71\" (UID: \"c2ca287e-027c-44b2-8068-b84437de7e71\") " Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.681661 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca" (OuterVolumeSpecName: "client-ca") pod "c2ca287e-027c-44b2-8068-b84437de7e71" (UID: "c2ca287e-027c-44b2-8068-b84437de7e71"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.681814 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config" (OuterVolumeSpecName: "config") pod "c2ca287e-027c-44b2-8068-b84437de7e71" (UID: "c2ca287e-027c-44b2-8068-b84437de7e71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.684425 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c2ca287e-027c-44b2-8068-b84437de7e71" (UID: "c2ca287e-027c-44b2-8068-b84437de7e71"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.685699 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m" (OuterVolumeSpecName: "kube-api-access-v6g7m") pod "c2ca287e-027c-44b2-8068-b84437de7e71" (UID: "c2ca287e-027c-44b2-8068-b84437de7e71"). InnerVolumeSpecName "kube-api-access-v6g7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.781074 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.781445 4726 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ca287e-027c-44b2-8068-b84437de7e71-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.781473 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6g7m\" (UniqueName: \"kubernetes.io/projected/c2ca287e-027c-44b2-8068-b84437de7e71-kube-api-access-v6g7m\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:58 crc kubenswrapper[4726]: I1004 03:52:58.781493 4726 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ca287e-027c-44b2-8068-b84437de7e71-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.218854 4726 scope.go:117] "RemoveContainer" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:52:59 crc kubenswrapper[4726]: E1004 03:52:59.219596 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\": container with ID starting with 18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7 not found: ID does not exist" containerID="18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.219672 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7"} err="failed to get container status \"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\": rpc error: code = NotFound desc = could not find container \"18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7\": container with ID starting with 18d0b6b6ef9a8fbc12ea49fea9b27e5e1a4863d88f8658e99c7179c49ae45eb7 not found: ID does not exist" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.219724 4726 scope.go:117] "RemoveContainer" containerID="3119c1176231bf658092c9b5347a02c4d01b5d6b25ef52b2592ada53e72bcee6" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.515683 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" event={"ID":"c2ca287e-027c-44b2-8068-b84437de7e71","Type":"ContainerDied","Data":"d4f66349559f4a30ec7a570ba461830dcd2cde794a40a1d6f26a9657e10412bb"} Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.515728 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.521459 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-snjd6_a8d6c450-3310-4058-90dc-0a39ab1934c4/ovn-controller/0.log" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.551507 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2"] Oct 04 03:52:59 crc kubenswrapper[4726]: E1004 03:52:59.551837 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.551861 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.552009 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" containerName="route-controller-manager" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.552692 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.555731 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.555833 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.555845 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.555885 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.556315 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.559218 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.564356 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.570440 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ttws2"] Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.593865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e9c5444-6e36-4816-8cf9-0086af3b739b-serving-cert\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.593930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-config\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.593972 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-client-ca\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.593996 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjglj\" (UniqueName: \"kubernetes.io/projected/4e9c5444-6e36-4816-8cf9-0086af3b739b-kube-api-access-sjglj\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.695059 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-client-ca\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.695142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjglj\" (UniqueName: \"kubernetes.io/projected/4e9c5444-6e36-4816-8cf9-0086af3b739b-kube-api-access-sjglj\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.695231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e9c5444-6e36-4816-8cf9-0086af3b739b-serving-cert\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.695272 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-config\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.696779 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-config\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.696906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e9c5444-6e36-4816-8cf9-0086af3b739b-client-ca\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.701618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e9c5444-6e36-4816-8cf9-0086af3b739b-serving-cert\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.715586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjglj\" (UniqueName: \"kubernetes.io/projected/4e9c5444-6e36-4816-8cf9-0086af3b739b-kube-api-access-sjglj\") pod \"route-controller-manager-6f7c945cc8-6cbh2\" (UID: \"4e9c5444-6e36-4816-8cf9-0086af3b739b\") " pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.870713 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.945698 4726 scope.go:117] "RemoveContainer" containerID="d4606dc9df927ffb361007695e2b0c1602e0ab9c09ae5e932e738449ac5ef71e" Oct 04 03:52:59 crc kubenswrapper[4726]: I1004 03:52:59.978358 4726 scope.go:117] "RemoveContainer" containerID="6fbf15e40a230f94d1029ce36928c46ee5888e02ba9c4f137555953f4cb94ee3" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.015423 4726 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(f13c1d0a210d5a714f8246297d9603fb3d8e7b70ceb5c1a6ba7804a6fc45394f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.015505 4726 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(f13c1d0a210d5a714f8246297d9603fb3d8e7b70ceb5c1a6ba7804a6fc45394f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.015530 4726 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(f13c1d0a210d5a714f8246297d9603fb3d8e7b70ceb5c1a6ba7804a6fc45394f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.015588 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-689c849788-ml4dx_openshift-controller-manager(d1beb981-cc80-4a6f-9ac3-cc837cc7b897)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-689c849788-ml4dx_openshift-controller-manager(d1beb981-cc80-4a6f-9ac3-cc837cc7b897)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(f13c1d0a210d5a714f8246297d9603fb3d8e7b70ceb5c1a6ba7804a6fc45394f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" podUID="d1beb981-cc80-4a6f-9ac3-cc837cc7b897" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.046482 4726 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(a29952fc7904237cde63d45b0360cbd6e4bafa154c78371a550afd29302614f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.046573 4726 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(a29952fc7904237cde63d45b0360cbd6e4bafa154c78371a550afd29302614f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.046616 4726 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(a29952fc7904237cde63d45b0360cbd6e4bafa154c78371a550afd29302614f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:00 crc kubenswrapper[4726]: E1004 03:53:00.046702 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager(4e9c5444-6e36-4816-8cf9-0086af3b739b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager(4e9c5444-6e36-4816-8cf9-0086af3b739b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(a29952fc7904237cde63d45b0360cbd6e4bafa154c78371a550afd29302614f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" podUID="4e9c5444-6e36-4816-8cf9-0086af3b739b" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.073004 4726 scope.go:117] "RemoveContainer" containerID="8f25701b9ecc4e66b40dc7ac624d06ead4578c473475b2ff20d143a835938bea" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.092208 4726 scope.go:117] "RemoveContainer" containerID="751f7bc7ea38973327e09dd5b37dfdc0a7082c3ae1dd082fb2faed8f114ba85f" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.117169 4726 scope.go:117] "RemoveContainer" containerID="7d266d270b02ff53f13b2976c7c44483e3b18cbf4b3b8752b97fdcfebbe9dde5" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.149210 4726 scope.go:117] "RemoveContainer" containerID="2749fbdebf2443e83ecdae267168238c8bd769f545c9162474473c6f368ff1cb" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.166161 4726 scope.go:117] "RemoveContainer" containerID="b63e43e0f1dc7a8da3ae311ee68f0c341345269bac03cde982719d79da1a9013" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.182563 4726 scope.go:117] "RemoveContainer" containerID="d90f0bb0450edf32ee4ebcc8370e81d6b84509f0f507f4bec08d62d2b3c8eff7" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.204088 4726 scope.go:117] "RemoveContainer" containerID="5e78fa72a941428ddbc806f3eebc4904263cb430ef50f4cb7a4cbefeb19b9562" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.509709 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0854df0d-1ca6-41a4-ba86-98fdec338fc1" path="/var/lib/kubelet/pods/0854df0d-1ca6-41a4-ba86-98fdec338fc1/volumes" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.511043 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ca287e-027c-44b2-8068-b84437de7e71" path="/var/lib/kubelet/pods/c2ca287e-027c-44b2-8068-b84437de7e71/volumes" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.529914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" event={"ID":"3628e5db-697e-495c-815f-c08b5ccd03da","Type":"ContainerStarted","Data":"9bb51467e9c8896a4693c24ca0d99e4b8b414b80c17c4d1ca85b5fc3689df0cd"} Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.532663 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vbkwl" event={"ID":"53efe0a4-7f9a-4541-9677-12e0dcec292e","Type":"ContainerStarted","Data":"e46bfce7307765d418ada657e2dab5d383a75a5c5d4754fc6e9d356ad3d9b394"} Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.535705 4726 generic.go:334] "Generic (PLEG): container finished" podID="fad4d495-d9e2-4428-a929-c9816972ac31" containerID="3978d48de059931eae5b87374641f4d5a2882777d2457e0891e12462b06a2e55" exitCode=0 Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.535772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerDied","Data":"3978d48de059931eae5b87374641f4d5a2882777d2457e0891e12462b06a2e55"} Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.551826 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-hkk79" podStartSLOduration=1.940832827 podStartE2EDuration="32.55180985s" podCreationTimestamp="2025-10-04 03:52:28 +0000 UTC" firstStartedPulling="2025-10-04 03:52:29.413573617 +0000 UTC m=+723.588196830" lastFinishedPulling="2025-10-04 03:53:00.02455063 +0000 UTC m=+754.199173853" observedRunningTime="2025-10-04 03:53:00.549718214 +0000 UTC m=+754.724341427" watchObservedRunningTime="2025-10-04 03:53:00.55180985 +0000 UTC m=+754.726433063" Oct 04 03:53:00 crc kubenswrapper[4726]: I1004 03:53:00.563095 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vbkwl" podStartSLOduration=1.987782212 podStartE2EDuration="32.563075179s" podCreationTimestamp="2025-10-04 03:52:28 +0000 UTC" firstStartedPulling="2025-10-04 03:52:29.449247413 +0000 UTC m=+723.623870626" lastFinishedPulling="2025-10-04 03:53:00.02454036 +0000 UTC m=+754.199163593" observedRunningTime="2025-10-04 03:53:00.563001627 +0000 UTC m=+754.737624840" watchObservedRunningTime="2025-10-04 03:53:00.563075179 +0000 UTC m=+754.737698392" Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548080 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"33161c3aee17364ca55d5f12559b97de21560733c4650c0e72753fbe47109482"} Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548614 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"83b359ae648b47e228dd592237427146be4666c49e7be627ae431cc8e4d40e42"} Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"2323d1ed390d85029501b4106eced8ca78c4953382f1b90f0b6b3a9eb42197b9"} Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"89d96340115b64809345352b650820d1acdce2155ba83549fa10579e224cb1d5"} Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548672 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"cc67ce02a4c11344206a14b779bb9f76a45e80a140e68a31ed0a385d5376a75b"} Oct 04 03:53:01 crc kubenswrapper[4726]: I1004 03:53:01.548690 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"d43316f38d3538fc64d1a60e96bf0268264a6f4e5b95e821ddc7d50da97c34cf"} Oct 04 03:53:03 crc kubenswrapper[4726]: I1004 03:53:03.205953 4726 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 03:53:04 crc kubenswrapper[4726]: I1004 03:53:04.188371 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:53:04 crc kubenswrapper[4726]: I1004 03:53:04.188747 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:53:04 crc kubenswrapper[4726]: I1004 03:53:04.573606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"5175e189b2aa412c8b65fb85e007afa58d3202993578b45950a36ee4aece3e37"} Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.604463 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" event={"ID":"fad4d495-d9e2-4428-a929-c9816972ac31","Type":"ContainerStarted","Data":"41ece3c5a2bf0021d1470443b68dedc6b19d82d48293864d955c102abc124ae1"} Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.605002 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.605014 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.632857 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.637904 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" podStartSLOduration=15.637863518 podStartE2EDuration="15.637863518s" podCreationTimestamp="2025-10-04 03:52:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:53:06.631978298 +0000 UTC m=+760.806601511" watchObservedRunningTime="2025-10-04 03:53:06.637863518 +0000 UTC m=+760.812486731" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.855235 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2"] Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.855363 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.855762 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.869376 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-689c849788-ml4dx"] Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.869493 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:06 crc kubenswrapper[4726]: I1004 03:53:06.869890 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.890081 4726 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(daa7081948a2b4385e2961bbc9e381b1d871c91f36958ef93a7db438541e1045): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.890178 4726 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(daa7081948a2b4385e2961bbc9e381b1d871c91f36958ef93a7db438541e1045): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.890212 4726 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(daa7081948a2b4385e2961bbc9e381b1d871c91f36958ef93a7db438541e1045): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.890272 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager(4e9c5444-6e36-4816-8cf9-0086af3b739b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager(4e9c5444-6e36-4816-8cf9-0086af3b739b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-6f7c945cc8-6cbh2_openshift-route-controller-manager_4e9c5444-6e36-4816-8cf9-0086af3b739b_0(daa7081948a2b4385e2961bbc9e381b1d871c91f36958ef93a7db438541e1045): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" podUID="4e9c5444-6e36-4816-8cf9-0086af3b739b" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.902774 4726 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(ee504175135e18c817a17ecdd7142f8c6df5807ecef6c0a2165f5471cabb17d1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.902840 4726 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(ee504175135e18c817a17ecdd7142f8c6df5807ecef6c0a2165f5471cabb17d1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.902864 4726 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(ee504175135e18c817a17ecdd7142f8c6df5807ecef6c0a2165f5471cabb17d1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:06 crc kubenswrapper[4726]: E1004 03:53:06.902915 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-689c849788-ml4dx_openshift-controller-manager(d1beb981-cc80-4a6f-9ac3-cc837cc7b897)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-689c849788-ml4dx_openshift-controller-manager(d1beb981-cc80-4a6f-9ac3-cc837cc7b897)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-689c849788-ml4dx_openshift-controller-manager_d1beb981-cc80-4a6f-9ac3-cc837cc7b897_0(ee504175135e18c817a17ecdd7142f8c6df5807ecef6c0a2165f5471cabb17d1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" podUID="d1beb981-cc80-4a6f-9ac3-cc837cc7b897" Oct 04 03:53:07 crc kubenswrapper[4726]: I1004 03:53:07.610926 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:07 crc kubenswrapper[4726]: I1004 03:53:07.647086 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:12 crc kubenswrapper[4726]: I1004 03:53:12.651954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" event={"ID":"5e7b44b7-601a-4d34-a2f5-13662f3ef3cd","Type":"ContainerStarted","Data":"023aa4e116e57056741030d31b5ca2a4c2ed8e4a002ece8d590b0b7b7793c972"} Oct 04 03:53:12 crc kubenswrapper[4726]: I1004 03:53:12.652914 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:53:12 crc kubenswrapper[4726]: I1004 03:53:12.673278 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" podStartSLOduration=1.8386847560000001 podStartE2EDuration="44.67325396s" podCreationTimestamp="2025-10-04 03:52:28 +0000 UTC" firstStartedPulling="2025-10-04 03:52:29.490673196 +0000 UTC m=+723.665296409" lastFinishedPulling="2025-10-04 03:53:12.32524236 +0000 UTC m=+766.499865613" observedRunningTime="2025-10-04 03:53:12.672617116 +0000 UTC m=+766.847240369" watchObservedRunningTime="2025-10-04 03:53:12.67325396 +0000 UTC m=+766.847877193" Oct 04 03:53:19 crc kubenswrapper[4726]: I1004 03:53:19.258019 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-nkt2j" Oct 04 03:53:19 crc kubenswrapper[4726]: I1004 03:53:19.501842 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:19 crc kubenswrapper[4726]: I1004 03:53:19.502267 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:19 crc kubenswrapper[4726]: I1004 03:53:19.748657 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-689c849788-ml4dx"] Oct 04 03:53:20 crc kubenswrapper[4726]: I1004 03:53:20.710114 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" event={"ID":"d1beb981-cc80-4a6f-9ac3-cc837cc7b897","Type":"ContainerStarted","Data":"22408fe5e08621fd8b37dffc6a84839a85c29a07e1af57399f6b711d681d1771"} Oct 04 03:53:20 crc kubenswrapper[4726]: I1004 03:53:20.710622 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" event={"ID":"d1beb981-cc80-4a6f-9ac3-cc837cc7b897","Type":"ContainerStarted","Data":"c88ef18bd961aeb711c6376771c07d843a9ff57bd199b0c5e5d0caf5cdbe21aa"} Oct 04 03:53:20 crc kubenswrapper[4726]: I1004 03:53:20.711948 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:20 crc kubenswrapper[4726]: I1004 03:53:20.721228 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" Oct 04 03:53:20 crc kubenswrapper[4726]: I1004 03:53:20.732034 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-689c849788-ml4dx" podStartSLOduration=24.732011726 podStartE2EDuration="24.732011726s" podCreationTimestamp="2025-10-04 03:52:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:53:20.729238254 +0000 UTC m=+774.903861467" watchObservedRunningTime="2025-10-04 03:53:20.732011726 +0000 UTC m=+774.906634939" Oct 04 03:53:21 crc kubenswrapper[4726]: I1004 03:53:21.974719 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2ssfl" Oct 04 03:53:22 crc kubenswrapper[4726]: I1004 03:53:22.501902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:22 crc kubenswrapper[4726]: I1004 03:53:22.502691 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:22 crc kubenswrapper[4726]: I1004 03:53:22.920174 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2"] Oct 04 03:53:22 crc kubenswrapper[4726]: W1004 03:53:22.932794 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e9c5444_6e36_4816_8cf9_0086af3b739b.slice/crio-39d081a1479a373530dde7a540d55eb514ce9316d853506c0c0c4b23078d7601 WatchSource:0}: Error finding container 39d081a1479a373530dde7a540d55eb514ce9316d853506c0c0c4b23078d7601: Status 404 returned error can't find the container with id 39d081a1479a373530dde7a540d55eb514ce9316d853506c0c0c4b23078d7601 Oct 04 03:53:23 crc kubenswrapper[4726]: I1004 03:53:23.732559 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" event={"ID":"4e9c5444-6e36-4816-8cf9-0086af3b739b","Type":"ContainerStarted","Data":"f2b60fc43ec1c0064b82e1b98aa83a8ee2b7627a010a94f178dbefe375dca614"} Oct 04 03:53:23 crc kubenswrapper[4726]: I1004 03:53:23.733333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" event={"ID":"4e9c5444-6e36-4816-8cf9-0086af3b739b","Type":"ContainerStarted","Data":"39d081a1479a373530dde7a540d55eb514ce9316d853506c0c0c4b23078d7601"} Oct 04 03:53:23 crc kubenswrapper[4726]: I1004 03:53:23.733408 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:23 crc kubenswrapper[4726]: I1004 03:53:23.742956 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" Oct 04 03:53:23 crc kubenswrapper[4726]: I1004 03:53:23.752487 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f7c945cc8-6cbh2" podStartSLOduration=26.752440694 podStartE2EDuration="26.752440694s" podCreationTimestamp="2025-10-04 03:52:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:53:23.750529456 +0000 UTC m=+777.925152669" watchObservedRunningTime="2025-10-04 03:53:23.752440694 +0000 UTC m=+777.927063977" Oct 04 03:53:29 crc kubenswrapper[4726]: I1004 03:53:29.877970 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:29 crc kubenswrapper[4726]: I1004 03:53:29.881782 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:29 crc kubenswrapper[4726]: I1004 03:53:29.906422 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.035095 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.035198 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrlfh\" (UniqueName: \"kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.035236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.136976 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.137056 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrlfh\" (UniqueName: \"kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.137091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.137685 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.137854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.156870 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrlfh\" (UniqueName: \"kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh\") pod \"community-operators-dkbwg\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.206318 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.663397 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:30 crc kubenswrapper[4726]: W1004 03:53:30.676900 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9539ee6c_bd42_4243_bcc7_7b1caf7243dc.slice/crio-bf03180176cbd4de8bda864b7e0fb41cc3458b6d1812daf743ce348e9bc28cc2 WatchSource:0}: Error finding container bf03180176cbd4de8bda864b7e0fb41cc3458b6d1812daf743ce348e9bc28cc2: Status 404 returned error can't find the container with id bf03180176cbd4de8bda864b7e0fb41cc3458b6d1812daf743ce348e9bc28cc2 Oct 04 03:53:30 crc kubenswrapper[4726]: I1004 03:53:30.778574 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerStarted","Data":"bf03180176cbd4de8bda864b7e0fb41cc3458b6d1812daf743ce348e9bc28cc2"} Oct 04 03:53:31 crc kubenswrapper[4726]: I1004 03:53:31.788529 4726 generic.go:334] "Generic (PLEG): container finished" podID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerID="bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7" exitCode=0 Oct 04 03:53:31 crc kubenswrapper[4726]: I1004 03:53:31.788586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerDied","Data":"bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7"} Oct 04 03:53:32 crc kubenswrapper[4726]: I1004 03:53:32.799859 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerStarted","Data":"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1"} Oct 04 03:53:33 crc kubenswrapper[4726]: I1004 03:53:33.813043 4726 generic.go:334] "Generic (PLEG): container finished" podID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerID="b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1" exitCode=0 Oct 04 03:53:33 crc kubenswrapper[4726]: I1004 03:53:33.813268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerDied","Data":"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1"} Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.188955 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.189051 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.189139 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.189960 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.190089 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db" gracePeriod=600 Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.825836 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerStarted","Data":"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939"} Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.831464 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db" exitCode=0 Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.831514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db"} Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.831544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7"} Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.831564 4726 scope.go:117] "RemoveContainer" containerID="4aa5ea00b4b8e6307c4a50e36b0ac7581ce7538e35d6eea411e96938916b7f41" Oct 04 03:53:34 crc kubenswrapper[4726]: I1004 03:53:34.848725 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dkbwg" podStartSLOduration=3.31808072 podStartE2EDuration="5.848695166s" podCreationTimestamp="2025-10-04 03:53:29 +0000 UTC" firstStartedPulling="2025-10-04 03:53:31.792010318 +0000 UTC m=+785.966633521" lastFinishedPulling="2025-10-04 03:53:34.322624714 +0000 UTC m=+788.497247967" observedRunningTime="2025-10-04 03:53:34.843488353 +0000 UTC m=+789.018111576" watchObservedRunningTime="2025-10-04 03:53:34.848695166 +0000 UTC m=+789.023318419" Oct 04 03:53:40 crc kubenswrapper[4726]: I1004 03:53:40.206662 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:40 crc kubenswrapper[4726]: I1004 03:53:40.207452 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:40 crc kubenswrapper[4726]: I1004 03:53:40.257542 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:40 crc kubenswrapper[4726]: I1004 03:53:40.938561 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:40 crc kubenswrapper[4726]: I1004 03:53:40.993577 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:42 crc kubenswrapper[4726]: I1004 03:53:42.885359 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dkbwg" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="registry-server" containerID="cri-o://2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939" gracePeriod=2 Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.441836 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.626910 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content\") pod \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.627239 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrlfh\" (UniqueName: \"kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh\") pod \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.627307 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities\") pod \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\" (UID: \"9539ee6c-bd42-4243-bcc7-7b1caf7243dc\") " Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.628826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities" (OuterVolumeSpecName: "utilities") pod "9539ee6c-bd42-4243-bcc7-7b1caf7243dc" (UID: "9539ee6c-bd42-4243-bcc7-7b1caf7243dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.635961 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh" (OuterVolumeSpecName: "kube-api-access-xrlfh") pod "9539ee6c-bd42-4243-bcc7-7b1caf7243dc" (UID: "9539ee6c-bd42-4243-bcc7-7b1caf7243dc"). InnerVolumeSpecName "kube-api-access-xrlfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.726409 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9539ee6c-bd42-4243-bcc7-7b1caf7243dc" (UID: "9539ee6c-bd42-4243-bcc7-7b1caf7243dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.729619 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrlfh\" (UniqueName: \"kubernetes.io/projected/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-kube-api-access-xrlfh\") on node \"crc\" DevicePath \"\"" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.729689 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.729716 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9539ee6c-bd42-4243-bcc7-7b1caf7243dc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.895582 4726 generic.go:334] "Generic (PLEG): container finished" podID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerID="2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939" exitCode=0 Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.895650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerDied","Data":"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939"} Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.895699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dkbwg" event={"ID":"9539ee6c-bd42-4243-bcc7-7b1caf7243dc","Type":"ContainerDied","Data":"bf03180176cbd4de8bda864b7e0fb41cc3458b6d1812daf743ce348e9bc28cc2"} Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.895730 4726 scope.go:117] "RemoveContainer" containerID="2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.895662 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dkbwg" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.929896 4726 scope.go:117] "RemoveContainer" containerID="b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.938138 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.942206 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dkbwg"] Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.956209 4726 scope.go:117] "RemoveContainer" containerID="bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.971392 4726 scope.go:117] "RemoveContainer" containerID="2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939" Oct 04 03:53:43 crc kubenswrapper[4726]: E1004 03:53:43.971959 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939\": container with ID starting with 2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939 not found: ID does not exist" containerID="2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.972014 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939"} err="failed to get container status \"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939\": rpc error: code = NotFound desc = could not find container \"2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939\": container with ID starting with 2477634ca4c89a2f020e3e18ba77bba879233b83156a72cc0e0c4af1851df939 not found: ID does not exist" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.972047 4726 scope.go:117] "RemoveContainer" containerID="b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1" Oct 04 03:53:43 crc kubenswrapper[4726]: E1004 03:53:43.972480 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1\": container with ID starting with b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1 not found: ID does not exist" containerID="b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.972521 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1"} err="failed to get container status \"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1\": rpc error: code = NotFound desc = could not find container \"b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1\": container with ID starting with b5d7f38a8e5c9b1965b0a32d9f7c8f8024e2b939e57f65b10950851eb216a2c1 not found: ID does not exist" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.972549 4726 scope.go:117] "RemoveContainer" containerID="bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7" Oct 04 03:53:43 crc kubenswrapper[4726]: E1004 03:53:43.972899 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7\": container with ID starting with bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7 not found: ID does not exist" containerID="bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7" Oct 04 03:53:43 crc kubenswrapper[4726]: I1004 03:53:43.972965 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7"} err="failed to get container status \"bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7\": rpc error: code = NotFound desc = could not find container \"bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7\": container with ID starting with bf10c6db2dc1fab59b8a4c58c2be11f61150b36946a47f26ad93fc6659b142e7 not found: ID does not exist" Oct 04 03:53:44 crc kubenswrapper[4726]: I1004 03:53:44.511714 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" path="/var/lib/kubelet/pods/9539ee6c-bd42-4243-bcc7-7b1caf7243dc/volumes" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.747498 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:53:52 crc kubenswrapper[4726]: E1004 03:53:52.748290 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="extract-content" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.748307 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="extract-content" Oct 04 03:53:52 crc kubenswrapper[4726]: E1004 03:53:52.748317 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="extract-utilities" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.748326 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="extract-utilities" Oct 04 03:53:52 crc kubenswrapper[4726]: E1004 03:53:52.748359 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="registry-server" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.748367 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="registry-server" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.748473 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9539ee6c-bd42-4243-bcc7-7b1caf7243dc" containerName="registry-server" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.749284 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.761801 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.777506 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.778531 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5wng\" (UniqueName: \"kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.778694 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.879633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.879720 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.879756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5wng\" (UniqueName: \"kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.880657 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.881296 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:52 crc kubenswrapper[4726]: I1004 03:53:52.904024 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5wng\" (UniqueName: \"kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng\") pod \"certified-operators-t8vkg\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:53 crc kubenswrapper[4726]: I1004 03:53:53.082672 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:53:53 crc kubenswrapper[4726]: I1004 03:53:53.541251 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:53:53 crc kubenswrapper[4726]: W1004 03:53:53.549754 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd6ecf1d_3756_4223_a680_1ea7763226bb.slice/crio-839ab519250b90e055bc985c70e49806e8720983ffd9c5ca445a4288ed3a19f0 WatchSource:0}: Error finding container 839ab519250b90e055bc985c70e49806e8720983ffd9c5ca445a4288ed3a19f0: Status 404 returned error can't find the container with id 839ab519250b90e055bc985c70e49806e8720983ffd9c5ca445a4288ed3a19f0 Oct 04 03:53:53 crc kubenswrapper[4726]: I1004 03:53:53.974649 4726 generic.go:334] "Generic (PLEG): container finished" podID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerID="369037e88904beffbd3b869f08dd68235171e0fb5c128d874546416dfccebdf9" exitCode=0 Oct 04 03:53:53 crc kubenswrapper[4726]: I1004 03:53:53.974725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerDied","Data":"369037e88904beffbd3b869f08dd68235171e0fb5c128d874546416dfccebdf9"} Oct 04 03:53:53 crc kubenswrapper[4726]: I1004 03:53:53.974920 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerStarted","Data":"839ab519250b90e055bc985c70e49806e8720983ffd9c5ca445a4288ed3a19f0"} Oct 04 03:53:54 crc kubenswrapper[4726]: I1004 03:53:54.983183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerStarted","Data":"d82c7bd09a53c7405acdd621d3c0754e4bfffa42849a594ec7c559132815676b"} Oct 04 03:53:55 crc kubenswrapper[4726]: I1004 03:53:55.993471 4726 generic.go:334] "Generic (PLEG): container finished" podID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerID="d82c7bd09a53c7405acdd621d3c0754e4bfffa42849a594ec7c559132815676b" exitCode=0 Oct 04 03:53:55 crc kubenswrapper[4726]: I1004 03:53:55.994392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerDied","Data":"d82c7bd09a53c7405acdd621d3c0754e4bfffa42849a594ec7c559132815676b"} Oct 04 03:53:57 crc kubenswrapper[4726]: I1004 03:53:57.002196 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerStarted","Data":"f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd"} Oct 04 03:53:57 crc kubenswrapper[4726]: I1004 03:53:57.021864 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t8vkg" podStartSLOduration=2.602433382 podStartE2EDuration="5.021845517s" podCreationTimestamp="2025-10-04 03:53:52 +0000 UTC" firstStartedPulling="2025-10-04 03:53:53.976844445 +0000 UTC m=+808.151467688" lastFinishedPulling="2025-10-04 03:53:56.39625659 +0000 UTC m=+810.570879823" observedRunningTime="2025-10-04 03:53:57.021644042 +0000 UTC m=+811.196267255" watchObservedRunningTime="2025-10-04 03:53:57.021845517 +0000 UTC m=+811.196468730" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.398328 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5"] Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.399902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.413310 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5"] Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.415145 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.564988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.565067 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhbbk\" (UniqueName: \"kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.565191 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.666690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.666745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhbbk\" (UniqueName: \"kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.666871 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.667704 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.668067 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.699628 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhbbk\" (UniqueName: \"kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:53:59 crc kubenswrapper[4726]: I1004 03:53:59.727185 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:54:00 crc kubenswrapper[4726]: I1004 03:54:00.201044 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5"] Oct 04 03:54:00 crc kubenswrapper[4726]: W1004 03:54:00.209915 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafad42c8_8dff_4c05_a9b8_08f750cbedec.slice/crio-40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37 WatchSource:0}: Error finding container 40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37: Status 404 returned error can't find the container with id 40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37 Oct 04 03:54:01 crc kubenswrapper[4726]: I1004 03:54:01.028355 4726 generic.go:334] "Generic (PLEG): container finished" podID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerID="399e04058ee074edacf929a034076f83cd8294317bcc32993d99e5265adf589f" exitCode=0 Oct 04 03:54:01 crc kubenswrapper[4726]: I1004 03:54:01.028698 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" event={"ID":"afad42c8-8dff-4c05-a9b8-08f750cbedec","Type":"ContainerDied","Data":"399e04058ee074edacf929a034076f83cd8294317bcc32993d99e5265adf589f"} Oct 04 03:54:01 crc kubenswrapper[4726]: I1004 03:54:01.028737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" event={"ID":"afad42c8-8dff-4c05-a9b8-08f750cbedec","Type":"ContainerStarted","Data":"40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37"} Oct 04 03:54:02 crc kubenswrapper[4726]: I1004 03:54:02.949376 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:02 crc kubenswrapper[4726]: I1004 03:54:02.951715 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:02 crc kubenswrapper[4726]: I1004 03:54:02.964418 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.044545 4726 generic.go:334] "Generic (PLEG): container finished" podID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerID="3ca84df337a9825da6f28d63784bc7821355c172e6b2397b584a90ff6ee06492" exitCode=0 Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.044592 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" event={"ID":"afad42c8-8dff-4c05-a9b8-08f750cbedec","Type":"ContainerDied","Data":"3ca84df337a9825da6f28d63784bc7821355c172e6b2397b584a90ff6ee06492"} Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.083431 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.083504 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.118475 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.118560 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.118647 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vdr\" (UniqueName: \"kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.156567 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.220084 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vdr\" (UniqueName: \"kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.220213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.220286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.221331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.221376 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.259202 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vdr\" (UniqueName: \"kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr\") pod \"redhat-operators-6ztkx\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.283935 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:03 crc kubenswrapper[4726]: I1004 03:54:03.703265 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:03 crc kubenswrapper[4726]: W1004 03:54:03.712284 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92d8370c_e789_486d_ad58_39ba53ef10f9.slice/crio-9671a61947b26c0b426a51907eeadfc75cf1982fc811733ee3840ab88bd4a4f2 WatchSource:0}: Error finding container 9671a61947b26c0b426a51907eeadfc75cf1982fc811733ee3840ab88bd4a4f2: Status 404 returned error can't find the container with id 9671a61947b26c0b426a51907eeadfc75cf1982fc811733ee3840ab88bd4a4f2 Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.050407 4726 generic.go:334] "Generic (PLEG): container finished" podID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerID="44bf34d936af4165b40dfc1e8ebd408642a5188694880bc4eddf55df5f41c99a" exitCode=0 Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.050475 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" event={"ID":"afad42c8-8dff-4c05-a9b8-08f750cbedec","Type":"ContainerDied","Data":"44bf34d936af4165b40dfc1e8ebd408642a5188694880bc4eddf55df5f41c99a"} Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.051779 4726 generic.go:334] "Generic (PLEG): container finished" podID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerID="f45b3a0eb2b9a416acffc6c2406d8d08c9b609631edc7181beafd622e1eb8d85" exitCode=0 Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.051857 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerDied","Data":"f45b3a0eb2b9a416acffc6c2406d8d08c9b609631edc7181beafd622e1eb8d85"} Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.051908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerStarted","Data":"9671a61947b26c0b426a51907eeadfc75cf1982fc811733ee3840ab88bd4a4f2"} Oct 04 03:54:04 crc kubenswrapper[4726]: I1004 03:54:04.091986 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.402267 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.578099 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util\") pod \"afad42c8-8dff-4c05-a9b8-08f750cbedec\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.578274 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle\") pod \"afad42c8-8dff-4c05-a9b8-08f750cbedec\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.578935 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhbbk\" (UniqueName: \"kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk\") pod \"afad42c8-8dff-4c05-a9b8-08f750cbedec\" (UID: \"afad42c8-8dff-4c05-a9b8-08f750cbedec\") " Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.579086 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle" (OuterVolumeSpecName: "bundle") pod "afad42c8-8dff-4c05-a9b8-08f750cbedec" (UID: "afad42c8-8dff-4c05-a9b8-08f750cbedec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.579646 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.586638 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk" (OuterVolumeSpecName: "kube-api-access-dhbbk") pod "afad42c8-8dff-4c05-a9b8-08f750cbedec" (UID: "afad42c8-8dff-4c05-a9b8-08f750cbedec"). InnerVolumeSpecName "kube-api-access-dhbbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.595092 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util" (OuterVolumeSpecName: "util") pod "afad42c8-8dff-4c05-a9b8-08f750cbedec" (UID: "afad42c8-8dff-4c05-a9b8-08f750cbedec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.680300 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afad42c8-8dff-4c05-a9b8-08f750cbedec-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:05 crc kubenswrapper[4726]: I1004 03:54:05.680339 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhbbk\" (UniqueName: \"kubernetes.io/projected/afad42c8-8dff-4c05-a9b8-08f750cbedec-kube-api-access-dhbbk\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.069837 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" event={"ID":"afad42c8-8dff-4c05-a9b8-08f750cbedec","Type":"ContainerDied","Data":"40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37"} Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.069884 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40bf8b328dac00b70f66494b2fb3888c04f473579d04c2e91d9bffbf2ecc7e37" Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.069901 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5" Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.072756 4726 generic.go:334] "Generic (PLEG): container finished" podID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerID="61fe6aa5ed922684430bd7f46f8abcc227edc0c469044fd3ec108a27131edbbd" exitCode=0 Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.072821 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerDied","Data":"61fe6aa5ed922684430bd7f46f8abcc227edc0c469044fd3ec108a27131edbbd"} Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.941677 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:54:06 crc kubenswrapper[4726]: I1004 03:54:06.942038 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t8vkg" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="registry-server" containerID="cri-o://f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd" gracePeriod=2 Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.053695 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd6ecf1d_3756_4223_a680_1ea7763226bb.slice/crio-conmon-f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.080684 4726 generic.go:334] "Generic (PLEG): container finished" podID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerID="f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd" exitCode=0 Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.080752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerDied","Data":"f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd"} Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.083146 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerStarted","Data":"dbb5d7977aeb0b582cc878f3ca90ffd2cc42438b99b8bbed2d01e9257a697414"} Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.132266 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6ztkx" podStartSLOduration=2.430614932 podStartE2EDuration="5.132250766s" podCreationTimestamp="2025-10-04 03:54:02 +0000 UTC" firstStartedPulling="2025-10-04 03:54:04.053919618 +0000 UTC m=+818.228542821" lastFinishedPulling="2025-10-04 03:54:06.755555402 +0000 UTC m=+820.930178655" observedRunningTime="2025-10-04 03:54:07.131363873 +0000 UTC m=+821.305987096" watchObservedRunningTime="2025-10-04 03:54:07.132250766 +0000 UTC m=+821.306873979" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.374680 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.513183 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities\") pod \"fd6ecf1d-3756-4223-a680-1ea7763226bb\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.513266 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5wng\" (UniqueName: \"kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng\") pod \"fd6ecf1d-3756-4223-a680-1ea7763226bb\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.513340 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content\") pod \"fd6ecf1d-3756-4223-a680-1ea7763226bb\" (UID: \"fd6ecf1d-3756-4223-a680-1ea7763226bb\") " Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.514889 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities" (OuterVolumeSpecName: "utilities") pod "fd6ecf1d-3756-4223-a680-1ea7763226bb" (UID: "fd6ecf1d-3756-4223-a680-1ea7763226bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.522312 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng" (OuterVolumeSpecName: "kube-api-access-p5wng") pod "fd6ecf1d-3756-4223-a680-1ea7763226bb" (UID: "fd6ecf1d-3756-4223-a680-1ea7763226bb"). InnerVolumeSpecName "kube-api-access-p5wng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.568582 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd6ecf1d-3756-4223-a680-1ea7763226bb" (UID: "fd6ecf1d-3756-4223-a680-1ea7763226bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.614256 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.614292 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5wng\" (UniqueName: \"kubernetes.io/projected/fd6ecf1d-3756-4223-a680-1ea7763226bb-kube-api-access-p5wng\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.614307 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd6ecf1d-3756-4223-a680-1ea7763226bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643617 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-s4p69"] Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643864 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="registry-server" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643878 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="registry-server" Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643891 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="pull" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643900 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="pull" Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643916 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="extract-utilities" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643926 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="extract-utilities" Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643935 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="extract" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643943 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="extract" Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643955 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="util" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643963 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="util" Oct 04 03:54:07 crc kubenswrapper[4726]: E1004 03:54:07.643978 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="extract-content" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.643986 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="extract-content" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.644121 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" containerName="registry-server" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.644150 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="afad42c8-8dff-4c05-a9b8-08f750cbedec" containerName="extract" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.644606 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.647702 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.647910 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-gsfdx" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.648071 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.660203 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-s4p69"] Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.714884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjcnl\" (UniqueName: \"kubernetes.io/projected/a5ee8c2e-8f69-4301-8aab-c5af01402c8a-kube-api-access-jjcnl\") pod \"nmstate-operator-858ddd8f98-s4p69\" (UID: \"a5ee8c2e-8f69-4301-8aab-c5af01402c8a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.816642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjcnl\" (UniqueName: \"kubernetes.io/projected/a5ee8c2e-8f69-4301-8aab-c5af01402c8a-kube-api-access-jjcnl\") pod \"nmstate-operator-858ddd8f98-s4p69\" (UID: \"a5ee8c2e-8f69-4301-8aab-c5af01402c8a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.842139 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjcnl\" (UniqueName: \"kubernetes.io/projected/a5ee8c2e-8f69-4301-8aab-c5af01402c8a-kube-api-access-jjcnl\") pod \"nmstate-operator-858ddd8f98-s4p69\" (UID: \"a5ee8c2e-8f69-4301-8aab-c5af01402c8a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" Oct 04 03:54:07 crc kubenswrapper[4726]: I1004 03:54:07.968787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.095833 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8vkg" Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.096246 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8vkg" event={"ID":"fd6ecf1d-3756-4223-a680-1ea7763226bb","Type":"ContainerDied","Data":"839ab519250b90e055bc985c70e49806e8720983ffd9c5ca445a4288ed3a19f0"} Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.096281 4726 scope.go:117] "RemoveContainer" containerID="f74cf2c7e9f4193583c8b07e9d3296bd07d2981149ca294fd305a6b519c6d9cd" Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.129806 4726 scope.go:117] "RemoveContainer" containerID="d82c7bd09a53c7405acdd621d3c0754e4bfffa42849a594ec7c559132815676b" Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.131926 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.136178 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t8vkg"] Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.162299 4726 scope.go:117] "RemoveContainer" containerID="369037e88904beffbd3b869f08dd68235171e0fb5c128d874546416dfccebdf9" Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.371935 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-s4p69"] Oct 04 03:54:08 crc kubenswrapper[4726]: W1004 03:54:08.378940 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5ee8c2e_8f69_4301_8aab_c5af01402c8a.slice/crio-ca32bc75b833b7006d9ece406cb55473fa3ae2877e5b5e50234b3a3aa0722f9a WatchSource:0}: Error finding container ca32bc75b833b7006d9ece406cb55473fa3ae2877e5b5e50234b3a3aa0722f9a: Status 404 returned error can't find the container with id ca32bc75b833b7006d9ece406cb55473fa3ae2877e5b5e50234b3a3aa0722f9a Oct 04 03:54:08 crc kubenswrapper[4726]: I1004 03:54:08.508906 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd6ecf1d-3756-4223-a680-1ea7763226bb" path="/var/lib/kubelet/pods/fd6ecf1d-3756-4223-a680-1ea7763226bb/volumes" Oct 04 03:54:09 crc kubenswrapper[4726]: I1004 03:54:09.103347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" event={"ID":"a5ee8c2e-8f69-4301-8aab-c5af01402c8a","Type":"ContainerStarted","Data":"ca32bc75b833b7006d9ece406cb55473fa3ae2877e5b5e50234b3a3aa0722f9a"} Oct 04 03:54:12 crc kubenswrapper[4726]: I1004 03:54:12.126345 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" event={"ID":"a5ee8c2e-8f69-4301-8aab-c5af01402c8a","Type":"ContainerStarted","Data":"62d17c9d6ededa871e39934f546a1055bb194ebe68d526f2e8d556584497be96"} Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.224426 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-s4p69" podStartSLOduration=3.746785076 podStartE2EDuration="6.224407107s" podCreationTimestamp="2025-10-04 03:54:07 +0000 UTC" firstStartedPulling="2025-10-04 03:54:08.381471191 +0000 UTC m=+822.556094404" lastFinishedPulling="2025-10-04 03:54:10.859093222 +0000 UTC m=+825.033716435" observedRunningTime="2025-10-04 03:54:12.151442051 +0000 UTC m=+826.326065304" watchObservedRunningTime="2025-10-04 03:54:13.224407107 +0000 UTC m=+827.399030320" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.227146 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.228180 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.230474 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-f6dpm" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.237507 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.238317 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.239720 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.258392 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jx5pc"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.259563 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.267228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.284572 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.301712 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.303871 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.380534 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.380808 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.381491 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.383301 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-chkrb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.384283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.386232 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.386765 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409560 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b2974204-71e0-4cd1-859b-0aef97909ee9-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xcn\" (UniqueName: \"kubernetes.io/projected/10fdef1b-c247-4b30-82b9-cde45b292b17-kube-api-access-x4xcn\") pod \"nmstate-metrics-fdff9cb8d-8n4qb\" (UID: \"10fdef1b-c247-4b30-82b9-cde45b292b17\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-ovs-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409682 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-nmstate-lock\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409701 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkd4r\" (UniqueName: \"kubernetes.io/projected/b0606264-b345-49e2-9965-806462686e84-kube-api-access-hkd4r\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409757 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2xw\" (UniqueName: \"kubernetes.io/projected/b2974204-71e0-4cd1-859b-0aef97909ee9-kube-api-access-6x2xw\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.409772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-dbus-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510527 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afe634c-ed15-4abf-998a-5929ab4c94c3-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510576 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s75g\" (UniqueName: \"kubernetes.io/projected/0afe634c-ed15-4abf-998a-5929ab4c94c3-kube-api-access-6s75g\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510607 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b2974204-71e0-4cd1-859b-0aef97909ee9-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xcn\" (UniqueName: \"kubernetes.io/projected/10fdef1b-c247-4b30-82b9-cde45b292b17-kube-api-access-x4xcn\") pod \"nmstate-metrics-fdff9cb8d-8n4qb\" (UID: \"10fdef1b-c247-4b30-82b9-cde45b292b17\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510791 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-ovs-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510828 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afe634c-ed15-4abf-998a-5929ab4c94c3-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-nmstate-lock\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510897 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkd4r\" (UniqueName: \"kubernetes.io/projected/b0606264-b345-49e2-9965-806462686e84-kube-api-access-hkd4r\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510939 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-nmstate-lock\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2xw\" (UniqueName: \"kubernetes.io/projected/b2974204-71e0-4cd1-859b-0aef97909ee9-kube-api-access-6x2xw\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510990 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-dbus-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.510978 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-ovs-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.511208 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b0606264-b345-49e2-9965-806462686e84-dbus-socket\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.517692 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b2974204-71e0-4cd1-859b-0aef97909ee9-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.536890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkd4r\" (UniqueName: \"kubernetes.io/projected/b0606264-b345-49e2-9965-806462686e84-kube-api-access-hkd4r\") pod \"nmstate-handler-jx5pc\" (UID: \"b0606264-b345-49e2-9965-806462686e84\") " pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.537289 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xcn\" (UniqueName: \"kubernetes.io/projected/10fdef1b-c247-4b30-82b9-cde45b292b17-kube-api-access-x4xcn\") pod \"nmstate-metrics-fdff9cb8d-8n4qb\" (UID: \"10fdef1b-c247-4b30-82b9-cde45b292b17\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.541774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2xw\" (UniqueName: \"kubernetes.io/projected/b2974204-71e0-4cd1-859b-0aef97909ee9-kube-api-access-6x2xw\") pod \"nmstate-webhook-6cdbc54649-mxq57\" (UID: \"b2974204-71e0-4cd1-859b-0aef97909ee9\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.558520 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.585608 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7558cc6967-v5gj4"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.586462 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.599293 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7558cc6967-v5gj4"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.612462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afe634c-ed15-4abf-998a-5929ab4c94c3-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.612504 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s75g\" (UniqueName: \"kubernetes.io/projected/0afe634c-ed15-4abf-998a-5929ab4c94c3-kube-api-access-6s75g\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.612539 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afe634c-ed15-4abf-998a-5929ab4c94c3-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.613540 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0afe634c-ed15-4abf-998a-5929ab4c94c3-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.615948 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0afe634c-ed15-4abf-998a-5929ab4c94c3-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.634020 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s75g\" (UniqueName: \"kubernetes.io/projected/0afe634c-ed15-4abf-998a-5929ab4c94c3-kube-api-access-6s75g\") pod \"nmstate-console-plugin-6b874cbd85-2tqv5\" (UID: \"0afe634c-ed15-4abf-998a-5929ab4c94c3\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.634258 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.643116 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.697514 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714041 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-trusted-ca-bundle\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714142 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfns2\" (UniqueName: \"kubernetes.io/projected/6fb859b4-292f-4765-acdf-dbc83c83215d-kube-api-access-qfns2\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714197 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-oauth-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-oauth-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714245 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714417 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-console-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.714522 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-service-ca\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816245 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-service-ca\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816318 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-trusted-ca-bundle\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816365 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfns2\" (UniqueName: \"kubernetes.io/projected/6fb859b4-292f-4765-acdf-dbc83c83215d-kube-api-access-qfns2\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816389 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-oauth-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-oauth-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.816475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-console-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.817758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-service-ca\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.818060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-oauth-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.818737 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-console-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.818800 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fb859b4-292f-4765-acdf-dbc83c83215d-trusted-ca-bundle\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.820971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-serving-cert\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.822016 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6fb859b4-292f-4765-acdf-dbc83c83215d-console-oauth-config\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.836729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfns2\" (UniqueName: \"kubernetes.io/projected/6fb859b4-292f-4765-acdf-dbc83c83215d-kube-api-access-qfns2\") pod \"console-7558cc6967-v5gj4\" (UID: \"6fb859b4-292f-4765-acdf-dbc83c83215d\") " pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.939044 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.943783 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.945631 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.950226 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:13 crc kubenswrapper[4726]: I1004 03:54:13.984806 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb"] Oct 04 03:54:13 crc kubenswrapper[4726]: W1004 03:54:13.996719 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10fdef1b_c247_4b30_82b9_cde45b292b17.slice/crio-16fd9051bb8c85a729a3c6462a25e93d4a163520cb7384f81df108030c0ec595 WatchSource:0}: Error finding container 16fd9051bb8c85a729a3c6462a25e93d4a163520cb7384f81df108030c0ec595: Status 404 returned error can't find the container with id 16fd9051bb8c85a729a3c6462a25e93d4a163520cb7384f81df108030c0ec595 Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.048729 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57"] Oct 04 03:54:14 crc kubenswrapper[4726]: W1004 03:54:14.075729 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2974204_71e0_4cd1_859b_0aef97909ee9.slice/crio-47b16dec428810e76f08c569a611f6ad75a29eb141eb31b59c81aa37c1f5baeb WatchSource:0}: Error finding container 47b16dec428810e76f08c569a611f6ad75a29eb141eb31b59c81aa37c1f5baeb: Status 404 returned error can't find the container with id 47b16dec428810e76f08c569a611f6ad75a29eb141eb31b59c81aa37c1f5baeb Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.106925 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5"] Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.120516 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.120589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.120731 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9hml\" (UniqueName: \"kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.136019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" event={"ID":"0afe634c-ed15-4abf-998a-5929ab4c94c3","Type":"ContainerStarted","Data":"9de8b94c2312bd1fbe370b03839862bc8d4348cf0984f69fcba65997d3ae2c54"} Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.137409 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" event={"ID":"10fdef1b-c247-4b30-82b9-cde45b292b17","Type":"ContainerStarted","Data":"16fd9051bb8c85a729a3c6462a25e93d4a163520cb7384f81df108030c0ec595"} Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.138989 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" event={"ID":"b2974204-71e0-4cd1-859b-0aef97909ee9","Type":"ContainerStarted","Data":"47b16dec428810e76f08c569a611f6ad75a29eb141eb31b59c81aa37c1f5baeb"} Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.142152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jx5pc" event={"ID":"b0606264-b345-49e2-9965-806462686e84","Type":"ContainerStarted","Data":"3f76f7fe4ab96a5d0728114de180e398352ae231b9984c8af484eb43dc4a56b6"} Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.195665 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.222011 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.222095 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.222182 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9hml\" (UniqueName: \"kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.222653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.222654 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.239900 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9hml\" (UniqueName: \"kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml\") pod \"redhat-marketplace-lpqn4\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.267697 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.451260 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7558cc6967-v5gj4"] Oct 04 03:54:14 crc kubenswrapper[4726]: W1004 03:54:14.459882 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fb859b4_292f_4765_acdf_dbc83c83215d.slice/crio-0a22abfb1fc806b055579cbc47c05b0f44047af61c2bc4922eeaa06123eacb72 WatchSource:0}: Error finding container 0a22abfb1fc806b055579cbc47c05b0f44047af61c2bc4922eeaa06123eacb72: Status 404 returned error can't find the container with id 0a22abfb1fc806b055579cbc47c05b0f44047af61c2bc4922eeaa06123eacb72 Oct 04 03:54:14 crc kubenswrapper[4726]: I1004 03:54:14.745804 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.147629 4726 generic.go:334] "Generic (PLEG): container finished" podID="86315df9-33b2-4367-9b55-0332f9c0d007" containerID="f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60" exitCode=0 Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.147685 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerDied","Data":"f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60"} Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.147707 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerStarted","Data":"22d3b425d57e2c5be6126fcab6baba7b82c3188924c08743cc7e22c32c08ffbd"} Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.150480 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7558cc6967-v5gj4" event={"ID":"6fb859b4-292f-4765-acdf-dbc83c83215d","Type":"ContainerStarted","Data":"2509b3bf369c5d5e296eb3355e7daecc254ce7661ad5418450018d7fe53856a3"} Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.150529 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7558cc6967-v5gj4" event={"ID":"6fb859b4-292f-4765-acdf-dbc83c83215d","Type":"ContainerStarted","Data":"0a22abfb1fc806b055579cbc47c05b0f44047af61c2bc4922eeaa06123eacb72"} Oct 04 03:54:15 crc kubenswrapper[4726]: I1004 03:54:15.183282 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7558cc6967-v5gj4" podStartSLOduration=2.18326317 podStartE2EDuration="2.18326317s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:54:15.177019982 +0000 UTC m=+829.351643195" watchObservedRunningTime="2025-10-04 03:54:15.18326317 +0000 UTC m=+829.357886393" Oct 04 03:54:17 crc kubenswrapper[4726]: I1004 03:54:17.528483 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:17 crc kubenswrapper[4726]: I1004 03:54:17.529023 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6ztkx" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="registry-server" containerID="cri-o://dbb5d7977aeb0b582cc878f3ca90ffd2cc42438b99b8bbed2d01e9257a697414" gracePeriod=2 Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.170710 4726 generic.go:334] "Generic (PLEG): container finished" podID="86315df9-33b2-4367-9b55-0332f9c0d007" containerID="efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b" exitCode=0 Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.170786 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerDied","Data":"efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b"} Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.172867 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" event={"ID":"10fdef1b-c247-4b30-82b9-cde45b292b17","Type":"ContainerStarted","Data":"8ba83431860a3227e8e7a12760bf1308502695c46da688ba2136af08d09a424c"} Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.175065 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" event={"ID":"b2974204-71e0-4cd1-859b-0aef97909ee9","Type":"ContainerStarted","Data":"8e5c223e6117c9bc0341e96fc8899635afb281ee67b3089e11018c45a19dc726"} Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.176858 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jx5pc" event={"ID":"b0606264-b345-49e2-9965-806462686e84","Type":"ContainerStarted","Data":"0a9b3e35537b26fdceb6008967f2b2d72840938afae77dfea09a2dae062daeaa"} Oct 04 03:54:18 crc kubenswrapper[4726]: I1004 03:54:18.179865 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" event={"ID":"0afe634c-ed15-4abf-998a-5929ab4c94c3","Type":"ContainerStarted","Data":"09f3cef38cb4386366901c7a576744689133907cafb81cf0a379c8515a70a265"} Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.202987 4726 generic.go:334] "Generic (PLEG): container finished" podID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerID="dbb5d7977aeb0b582cc878f3ca90ffd2cc42438b99b8bbed2d01e9257a697414" exitCode=0 Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.203088 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerDied","Data":"dbb5d7977aeb0b582cc878f3ca90ffd2cc42438b99b8bbed2d01e9257a697414"} Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.206809 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerStarted","Data":"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642"} Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.207282 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.207374 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.245526 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-2tqv5" podStartSLOduration=3.534148783 podStartE2EDuration="6.245501954s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="2025-10-04 03:54:14.115365944 +0000 UTC m=+828.289989157" lastFinishedPulling="2025-10-04 03:54:16.826719105 +0000 UTC m=+831.001342328" observedRunningTime="2025-10-04 03:54:19.228950484 +0000 UTC m=+833.403573767" watchObservedRunningTime="2025-10-04 03:54:19.245501954 +0000 UTC m=+833.420125167" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.260507 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jx5pc" podStartSLOduration=3.115884574 podStartE2EDuration="6.260476303s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="2025-10-04 03:54:13.679246033 +0000 UTC m=+827.853869246" lastFinishedPulling="2025-10-04 03:54:16.823837752 +0000 UTC m=+830.998460975" observedRunningTime="2025-10-04 03:54:19.252091311 +0000 UTC m=+833.426714524" watchObservedRunningTime="2025-10-04 03:54:19.260476303 +0000 UTC m=+833.435099556" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.276303 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lpqn4" podStartSLOduration=2.72970983 podStartE2EDuration="6.276271524s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="2025-10-04 03:54:15.148750295 +0000 UTC m=+829.323373508" lastFinishedPulling="2025-10-04 03:54:18.695311989 +0000 UTC m=+832.869935202" observedRunningTime="2025-10-04 03:54:19.273644837 +0000 UTC m=+833.448268070" watchObservedRunningTime="2025-10-04 03:54:19.276271524 +0000 UTC m=+833.450894787" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.427925 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.451501 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" podStartSLOduration=3.701530729 podStartE2EDuration="6.451480078s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="2025-10-04 03:54:14.079523855 +0000 UTC m=+828.254147088" lastFinishedPulling="2025-10-04 03:54:16.829473214 +0000 UTC m=+831.004096437" observedRunningTime="2025-10-04 03:54:19.293176303 +0000 UTC m=+833.467799556" watchObservedRunningTime="2025-10-04 03:54:19.451480078 +0000 UTC m=+833.626103301" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.516869 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content\") pod \"92d8370c-e789-486d-ad58-39ba53ef10f9\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.517031 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vdr\" (UniqueName: \"kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr\") pod \"92d8370c-e789-486d-ad58-39ba53ef10f9\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.517068 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities\") pod \"92d8370c-e789-486d-ad58-39ba53ef10f9\" (UID: \"92d8370c-e789-486d-ad58-39ba53ef10f9\") " Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.518000 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities" (OuterVolumeSpecName: "utilities") pod "92d8370c-e789-486d-ad58-39ba53ef10f9" (UID: "92d8370c-e789-486d-ad58-39ba53ef10f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.523906 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr" (OuterVolumeSpecName: "kube-api-access-m7vdr") pod "92d8370c-e789-486d-ad58-39ba53ef10f9" (UID: "92d8370c-e789-486d-ad58-39ba53ef10f9"). InnerVolumeSpecName "kube-api-access-m7vdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.618607 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vdr\" (UniqueName: \"kubernetes.io/projected/92d8370c-e789-486d-ad58-39ba53ef10f9-kube-api-access-m7vdr\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.618647 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.626175 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92d8370c-e789-486d-ad58-39ba53ef10f9" (UID: "92d8370c-e789-486d-ad58-39ba53ef10f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:19 crc kubenswrapper[4726]: I1004 03:54:19.720551 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92d8370c-e789-486d-ad58-39ba53ef10f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.217336 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6ztkx" Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.217423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6ztkx" event={"ID":"92d8370c-e789-486d-ad58-39ba53ef10f9","Type":"ContainerDied","Data":"9671a61947b26c0b426a51907eeadfc75cf1982fc811733ee3840ab88bd4a4f2"} Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.217521 4726 scope.go:117] "RemoveContainer" containerID="dbb5d7977aeb0b582cc878f3ca90ffd2cc42438b99b8bbed2d01e9257a697414" Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.239873 4726 scope.go:117] "RemoveContainer" containerID="61fe6aa5ed922684430bd7f46f8abcc227edc0c469044fd3ec108a27131edbbd" Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.257520 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.269856 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6ztkx"] Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.289240 4726 scope.go:117] "RemoveContainer" containerID="f45b3a0eb2b9a416acffc6c2406d8d08c9b609631edc7181beafd622e1eb8d85" Oct 04 03:54:20 crc kubenswrapper[4726]: I1004 03:54:20.511466 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" path="/var/lib/kubelet/pods/92d8370c-e789-486d-ad58-39ba53ef10f9/volumes" Oct 04 03:54:22 crc kubenswrapper[4726]: I1004 03:54:22.235216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" event={"ID":"10fdef1b-c247-4b30-82b9-cde45b292b17","Type":"ContainerStarted","Data":"d2b8d4cfcd682b9be57ff5a97647641847e39792e7cac811d289165e4b1ab4af"} Oct 04 03:54:22 crc kubenswrapper[4726]: I1004 03:54:22.266156 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8n4qb" podStartSLOduration=1.969801376 podStartE2EDuration="9.266090367s" podCreationTimestamp="2025-10-04 03:54:13 +0000 UTC" firstStartedPulling="2025-10-04 03:54:14.003569609 +0000 UTC m=+828.178192842" lastFinishedPulling="2025-10-04 03:54:21.29985862 +0000 UTC m=+835.474481833" observedRunningTime="2025-10-04 03:54:22.26422622 +0000 UTC m=+836.438849503" watchObservedRunningTime="2025-10-04 03:54:22.266090367 +0000 UTC m=+836.440713610" Oct 04 03:54:23 crc kubenswrapper[4726]: I1004 03:54:23.684619 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jx5pc" Oct 04 03:54:23 crc kubenswrapper[4726]: I1004 03:54:23.941343 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:23 crc kubenswrapper[4726]: I1004 03:54:23.941431 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:23 crc kubenswrapper[4726]: I1004 03:54:23.948644 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:24 crc kubenswrapper[4726]: I1004 03:54:24.249785 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7558cc6967-v5gj4" Oct 04 03:54:24 crc kubenswrapper[4726]: I1004 03:54:24.268251 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:24 crc kubenswrapper[4726]: I1004 03:54:24.268340 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:24 crc kubenswrapper[4726]: I1004 03:54:24.298699 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:54:24 crc kubenswrapper[4726]: I1004 03:54:24.322893 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:25 crc kubenswrapper[4726]: I1004 03:54:25.317270 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:26 crc kubenswrapper[4726]: I1004 03:54:26.539621 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.267854 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lpqn4" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="registry-server" containerID="cri-o://b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642" gracePeriod=2 Oct 04 03:54:27 crc kubenswrapper[4726]: E1004 03:54:27.437006 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86315df9_33b2_4367_9b55_0332f9c0d007.slice/crio-conmon-b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86315df9_33b2_4367_9b55_0332f9c0d007.slice/crio-b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.763786 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.837308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content\") pod \"86315df9-33b2-4367-9b55-0332f9c0d007\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.837361 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities\") pod \"86315df9-33b2-4367-9b55-0332f9c0d007\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.837422 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9hml\" (UniqueName: \"kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml\") pod \"86315df9-33b2-4367-9b55-0332f9c0d007\" (UID: \"86315df9-33b2-4367-9b55-0332f9c0d007\") " Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.844074 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities" (OuterVolumeSpecName: "utilities") pod "86315df9-33b2-4367-9b55-0332f9c0d007" (UID: "86315df9-33b2-4367-9b55-0332f9c0d007"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.846408 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml" (OuterVolumeSpecName: "kube-api-access-x9hml") pod "86315df9-33b2-4367-9b55-0332f9c0d007" (UID: "86315df9-33b2-4367-9b55-0332f9c0d007"). InnerVolumeSpecName "kube-api-access-x9hml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.861948 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86315df9-33b2-4367-9b55-0332f9c0d007" (UID: "86315df9-33b2-4367-9b55-0332f9c0d007"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.939492 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9hml\" (UniqueName: \"kubernetes.io/projected/86315df9-33b2-4367-9b55-0332f9c0d007-kube-api-access-x9hml\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.939551 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:27 crc kubenswrapper[4726]: I1004 03:54:27.939573 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86315df9-33b2-4367-9b55-0332f9c0d007-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.277468 4726 generic.go:334] "Generic (PLEG): container finished" podID="86315df9-33b2-4367-9b55-0332f9c0d007" containerID="b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642" exitCode=0 Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.277516 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerDied","Data":"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642"} Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.277548 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpqn4" event={"ID":"86315df9-33b2-4367-9b55-0332f9c0d007","Type":"ContainerDied","Data":"22d3b425d57e2c5be6126fcab6baba7b82c3188924c08743cc7e22c32c08ffbd"} Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.277569 4726 scope.go:117] "RemoveContainer" containerID="b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.277600 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpqn4" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.299985 4726 scope.go:117] "RemoveContainer" containerID="efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.330354 4726 scope.go:117] "RemoveContainer" containerID="f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.333374 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.340812 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpqn4"] Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.411793 4726 scope.go:117] "RemoveContainer" containerID="b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642" Oct 04 03:54:28 crc kubenswrapper[4726]: E1004 03:54:28.412932 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642\": container with ID starting with b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642 not found: ID does not exist" containerID="b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.412988 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642"} err="failed to get container status \"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642\": rpc error: code = NotFound desc = could not find container \"b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642\": container with ID starting with b0228723f5d579a9f26b8864a9b079ff4315fe4ecbef57e2a49634a901d5f642 not found: ID does not exist" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.413022 4726 scope.go:117] "RemoveContainer" containerID="efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b" Oct 04 03:54:28 crc kubenswrapper[4726]: E1004 03:54:28.413630 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b\": container with ID starting with efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b not found: ID does not exist" containerID="efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.413665 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b"} err="failed to get container status \"efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b\": rpc error: code = NotFound desc = could not find container \"efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b\": container with ID starting with efce63eafffae43b781c40cb2d93eacd0585942c7696dcacfbdfec849ee1ca0b not found: ID does not exist" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.413694 4726 scope.go:117] "RemoveContainer" containerID="f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60" Oct 04 03:54:28 crc kubenswrapper[4726]: E1004 03:54:28.414230 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60\": container with ID starting with f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60 not found: ID does not exist" containerID="f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.414293 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60"} err="failed to get container status \"f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60\": rpc error: code = NotFound desc = could not find container \"f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60\": container with ID starting with f36edbf70889e020007da564c7817d6ba99806fb528dd1789bc49c5c3da29f60 not found: ID does not exist" Oct 04 03:54:28 crc kubenswrapper[4726]: I1004 03:54:28.509464 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" path="/var/lib/kubelet/pods/86315df9-33b2-4367-9b55-0332f9c0d007/volumes" Oct 04 03:54:33 crc kubenswrapper[4726]: I1004 03:54:33.643821 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-mxq57" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.091179 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p"] Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092146 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="extract-utilities" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092169 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="extract-utilities" Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092193 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="extract-content" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092206 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="extract-content" Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092238 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="extract-utilities" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092251 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="extract-utilities" Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092269 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092281 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092299 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092311 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: E1004 03:54:49.092328 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="extract-content" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092342 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="extract-content" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092570 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d8370c-e789-486d-ad58-39ba53ef10f9" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.092602 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="86315df9-33b2-4367-9b55-0332f9c0d007" containerName="registry-server" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.094102 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.098715 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.116799 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p"] Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.208000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.208087 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.208185 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgc92\" (UniqueName: \"kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.308997 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.309070 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgc92\" (UniqueName: \"kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.309195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.309934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.309944 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.341275 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgc92\" (UniqueName: \"kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.367908 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-x6s4v" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerName="console" containerID="cri-o://7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0" gracePeriod=15 Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.458900 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.797294 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p"] Oct 04 03:54:49 crc kubenswrapper[4726]: W1004 03:54:49.809964 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddecbcd72_7ac4_4806_9b47_9611634efc23.slice/crio-e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118 WatchSource:0}: Error finding container e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118: Status 404 returned error can't find the container with id e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118 Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.837542 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-x6s4v_8cdb834a-e95f-4532-bd81-207fa47529aa/console/0.log" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.837619 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.917673 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.917981 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.918025 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.918048 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.918152 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2fw8\" (UniqueName: \"kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.918179 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.918211 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle\") pod \"8cdb834a-e95f-4532-bd81-207fa47529aa\" (UID: \"8cdb834a-e95f-4532-bd81-207fa47529aa\") " Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.919296 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca" (OuterVolumeSpecName: "service-ca") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.919340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.919401 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config" (OuterVolumeSpecName: "console-config") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.919322 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.924903 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.925087 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:54:49 crc kubenswrapper[4726]: I1004 03:54:49.925087 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8" (OuterVolumeSpecName: "kube-api-access-h2fw8") pod "8cdb834a-e95f-4532-bd81-207fa47529aa" (UID: "8cdb834a-e95f-4532-bd81-207fa47529aa"). InnerVolumeSpecName "kube-api-access-h2fw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020201 4726 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020254 4726 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020272 4726 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020288 4726 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020305 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2fw8\" (UniqueName: \"kubernetes.io/projected/8cdb834a-e95f-4532-bd81-207fa47529aa-kube-api-access-h2fw8\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020324 4726 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8cdb834a-e95f-4532-bd81-207fa47529aa-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.020354 4726 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8cdb834a-e95f-4532-bd81-207fa47529aa-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447077 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-x6s4v_8cdb834a-e95f-4532-bd81-207fa47529aa/console/0.log" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447189 4726 generic.go:334] "Generic (PLEG): container finished" podID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerID="7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0" exitCode=2 Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447330 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-x6s4v" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447295 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-x6s4v" event={"ID":"8cdb834a-e95f-4532-bd81-207fa47529aa","Type":"ContainerDied","Data":"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0"} Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447424 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-x6s4v" event={"ID":"8cdb834a-e95f-4532-bd81-207fa47529aa","Type":"ContainerDied","Data":"d639cbcef8aebbbbe7e1efe3a6a3f117e4f8354ba4f870f205b42dcbba66d3f3"} Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.447450 4726 scope.go:117] "RemoveContainer" containerID="7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.454259 4726 generic.go:334] "Generic (PLEG): container finished" podID="decbcd72-7ac4-4806-9b47-9611634efc23" containerID="ab79db1bb417149a277f91c36b6e351621ba2e59a8b6e4ac6a7ff9a9c48bee53" exitCode=0 Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.454321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" event={"ID":"decbcd72-7ac4-4806-9b47-9611634efc23","Type":"ContainerDied","Data":"ab79db1bb417149a277f91c36b6e351621ba2e59a8b6e4ac6a7ff9a9c48bee53"} Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.454370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" event={"ID":"decbcd72-7ac4-4806-9b47-9611634efc23","Type":"ContainerStarted","Data":"e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118"} Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.487567 4726 scope.go:117] "RemoveContainer" containerID="7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0" Oct 04 03:54:50 crc kubenswrapper[4726]: E1004 03:54:50.493484 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0\": container with ID starting with 7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0 not found: ID does not exist" containerID="7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.493530 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0"} err="failed to get container status \"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0\": rpc error: code = NotFound desc = could not find container \"7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0\": container with ID starting with 7be66e453807ad81680e996be085209b33f541ce4fcf861797f58eaf25f157c0 not found: ID does not exist" Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.495660 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.501404 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-x6s4v"] Oct 04 03:54:50 crc kubenswrapper[4726]: I1004 03:54:50.513926 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" path="/var/lib/kubelet/pods/8cdb834a-e95f-4532-bd81-207fa47529aa/volumes" Oct 04 03:54:53 crc kubenswrapper[4726]: I1004 03:54:53.484597 4726 generic.go:334] "Generic (PLEG): container finished" podID="decbcd72-7ac4-4806-9b47-9611634efc23" containerID="e6ca4928cee105aa41e78d34c6377d908f6efec35bf5686d44610da268c693cd" exitCode=0 Oct 04 03:54:53 crc kubenswrapper[4726]: I1004 03:54:53.484699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" event={"ID":"decbcd72-7ac4-4806-9b47-9611634efc23","Type":"ContainerDied","Data":"e6ca4928cee105aa41e78d34c6377d908f6efec35bf5686d44610da268c693cd"} Oct 04 03:54:54 crc kubenswrapper[4726]: I1004 03:54:54.496334 4726 generic.go:334] "Generic (PLEG): container finished" podID="decbcd72-7ac4-4806-9b47-9611634efc23" containerID="b4ff7949266798016771beefb32039f05fec5b5884f6e74e06f102c673bb0668" exitCode=0 Oct 04 03:54:54 crc kubenswrapper[4726]: I1004 03:54:54.496418 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" event={"ID":"decbcd72-7ac4-4806-9b47-9611634efc23","Type":"ContainerDied","Data":"b4ff7949266798016771beefb32039f05fec5b5884f6e74e06f102c673bb0668"} Oct 04 03:54:55 crc kubenswrapper[4726]: I1004 03:54:55.897603 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.001342 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgc92\" (UniqueName: \"kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92\") pod \"decbcd72-7ac4-4806-9b47-9611634efc23\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.001447 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util\") pod \"decbcd72-7ac4-4806-9b47-9611634efc23\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.001501 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle\") pod \"decbcd72-7ac4-4806-9b47-9611634efc23\" (UID: \"decbcd72-7ac4-4806-9b47-9611634efc23\") " Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.003218 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle" (OuterVolumeSpecName: "bundle") pod "decbcd72-7ac4-4806-9b47-9611634efc23" (UID: "decbcd72-7ac4-4806-9b47-9611634efc23"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.009444 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92" (OuterVolumeSpecName: "kube-api-access-zgc92") pod "decbcd72-7ac4-4806-9b47-9611634efc23" (UID: "decbcd72-7ac4-4806-9b47-9611634efc23"). InnerVolumeSpecName "kube-api-access-zgc92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.018632 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util" (OuterVolumeSpecName: "util") pod "decbcd72-7ac4-4806-9b47-9611634efc23" (UID: "decbcd72-7ac4-4806-9b47-9611634efc23"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.103462 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.103518 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/decbcd72-7ac4-4806-9b47-9611634efc23-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.103538 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgc92\" (UniqueName: \"kubernetes.io/projected/decbcd72-7ac4-4806-9b47-9611634efc23-kube-api-access-zgc92\") on node \"crc\" DevicePath \"\"" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.517431 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.518307 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p" event={"ID":"decbcd72-7ac4-4806-9b47-9611634efc23","Type":"ContainerDied","Data":"e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118"} Oct 04 03:54:56 crc kubenswrapper[4726]: I1004 03:54:56.518369 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e298194aa91ca5cc65143a1166923a7acf5bbe9e671b315357c0d88633059118" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.230544 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh"] Oct 04 03:55:07 crc kubenswrapper[4726]: E1004 03:55:07.231575 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerName="console" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231596 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerName="console" Oct 04 03:55:07 crc kubenswrapper[4726]: E1004 03:55:07.231629 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="pull" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231637 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="pull" Oct 04 03:55:07 crc kubenswrapper[4726]: E1004 03:55:07.231654 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="extract" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231662 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="extract" Oct 04 03:55:07 crc kubenswrapper[4726]: E1004 03:55:07.231689 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="util" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231696 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="util" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231907 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="decbcd72-7ac4-4806-9b47-9611634efc23" containerName="extract" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.231928 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cdb834a-e95f-4532-bd81-207fa47529aa" containerName="console" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.232601 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.235517 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.235808 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.236117 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.243348 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.243980 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-dzmwz" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.246849 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh"] Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.364098 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-webhook-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.364773 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-apiservice-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.364971 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8gz9\" (UniqueName: \"kubernetes.io/projected/898913f8-1833-4c8f-8764-32c656d6caab-kube-api-access-c8gz9\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.440450 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69"] Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.441055 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.444845 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.444867 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6zn6g" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.445050 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.461950 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69"] Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.465527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-webhook-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.465568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-webhook-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.465589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-apiservice-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.465666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-apiservice-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.466565 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x54s5\" (UniqueName: \"kubernetes.io/projected/898ba0b6-6440-45be-8067-6719ead41cba-kube-api-access-x54s5\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.466600 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8gz9\" (UniqueName: \"kubernetes.io/projected/898913f8-1833-4c8f-8764-32c656d6caab-kube-api-access-c8gz9\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.473954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-webhook-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.484209 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8gz9\" (UniqueName: \"kubernetes.io/projected/898913f8-1833-4c8f-8764-32c656d6caab-kube-api-access-c8gz9\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.487683 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898913f8-1833-4c8f-8764-32c656d6caab-apiservice-cert\") pod \"metallb-operator-controller-manager-655cdc4dc4-v2hjh\" (UID: \"898913f8-1833-4c8f-8764-32c656d6caab\") " pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.565850 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.568090 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x54s5\" (UniqueName: \"kubernetes.io/projected/898ba0b6-6440-45be-8067-6719ead41cba-kube-api-access-x54s5\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.568186 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-webhook-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.568215 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-apiservice-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.571789 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-webhook-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.572475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/898ba0b6-6440-45be-8067-6719ead41cba-apiservice-cert\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.589863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x54s5\" (UniqueName: \"kubernetes.io/projected/898ba0b6-6440-45be-8067-6719ead41cba-kube-api-access-x54s5\") pod \"metallb-operator-webhook-server-66d48bf4d6-nkg69\" (UID: \"898ba0b6-6440-45be-8067-6719ead41cba\") " pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.756216 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:07 crc kubenswrapper[4726]: I1004 03:55:07.999879 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh"] Oct 04 03:55:08 crc kubenswrapper[4726]: W1004 03:55:08.009763 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod898913f8_1833_4c8f_8764_32c656d6caab.slice/crio-fdffe4af9a0bc4af8f304e69b0325abcbb1148e9e071160014a06807b78990e4 WatchSource:0}: Error finding container fdffe4af9a0bc4af8f304e69b0325abcbb1148e9e071160014a06807b78990e4: Status 404 returned error can't find the container with id fdffe4af9a0bc4af8f304e69b0325abcbb1148e9e071160014a06807b78990e4 Oct 04 03:55:08 crc kubenswrapper[4726]: I1004 03:55:08.188466 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69"] Oct 04 03:55:08 crc kubenswrapper[4726]: W1004 03:55:08.192680 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod898ba0b6_6440_45be_8067_6719ead41cba.slice/crio-f4a7717220ebb502d0600b899e27677869db7f8aa34f71615688332eec568045 WatchSource:0}: Error finding container f4a7717220ebb502d0600b899e27677869db7f8aa34f71615688332eec568045: Status 404 returned error can't find the container with id f4a7717220ebb502d0600b899e27677869db7f8aa34f71615688332eec568045 Oct 04 03:55:08 crc kubenswrapper[4726]: I1004 03:55:08.581958 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" event={"ID":"898ba0b6-6440-45be-8067-6719ead41cba","Type":"ContainerStarted","Data":"f4a7717220ebb502d0600b899e27677869db7f8aa34f71615688332eec568045"} Oct 04 03:55:08 crc kubenswrapper[4726]: I1004 03:55:08.583382 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" event={"ID":"898913f8-1833-4c8f-8764-32c656d6caab","Type":"ContainerStarted","Data":"fdffe4af9a0bc4af8f304e69b0325abcbb1148e9e071160014a06807b78990e4"} Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.630333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" event={"ID":"898ba0b6-6440-45be-8067-6719ead41cba","Type":"ContainerStarted","Data":"09fd16e68225eba3e50dc452db8a24875229e4c32a0a096eddf375bf30c90610"} Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.631029 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.633455 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" event={"ID":"898913f8-1833-4c8f-8764-32c656d6caab","Type":"ContainerStarted","Data":"9b224063d02c071fc6025571ba49106fbae8f4804af988a31f195f966aed0d2f"} Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.633633 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.655604 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" podStartSLOduration=2.256686726 podStartE2EDuration="7.65558197s" podCreationTimestamp="2025-10-04 03:55:07 +0000 UTC" firstStartedPulling="2025-10-04 03:55:08.195715403 +0000 UTC m=+882.370338616" lastFinishedPulling="2025-10-04 03:55:13.594610647 +0000 UTC m=+887.769233860" observedRunningTime="2025-10-04 03:55:14.651079962 +0000 UTC m=+888.825703205" watchObservedRunningTime="2025-10-04 03:55:14.65558197 +0000 UTC m=+888.830205183" Oct 04 03:55:14 crc kubenswrapper[4726]: I1004 03:55:14.702362 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" podStartSLOduration=2.146976356 podStartE2EDuration="7.702342013s" podCreationTimestamp="2025-10-04 03:55:07 +0000 UTC" firstStartedPulling="2025-10-04 03:55:08.011959886 +0000 UTC m=+882.186583099" lastFinishedPulling="2025-10-04 03:55:13.567325543 +0000 UTC m=+887.741948756" observedRunningTime="2025-10-04 03:55:14.697733453 +0000 UTC m=+888.872356666" watchObservedRunningTime="2025-10-04 03:55:14.702342013 +0000 UTC m=+888.876965236" Oct 04 03:55:27 crc kubenswrapper[4726]: I1004 03:55:27.761282 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-66d48bf4d6-nkg69" Oct 04 03:55:34 crc kubenswrapper[4726]: I1004 03:55:34.188909 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:55:34 crc kubenswrapper[4726]: I1004 03:55:34.189469 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:55:47 crc kubenswrapper[4726]: I1004 03:55:47.569916 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-655cdc4dc4-v2hjh" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.447705 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zktqf"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.452933 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.453235 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.454096 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.465425 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.465842 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.466001 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.466234 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-kxs7n" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.470150 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.540525 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-89wr2"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.541618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.543287 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.543531 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.543742 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p9s5l" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.546355 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqbgf\" (UniqueName: \"kubernetes.io/projected/555989d8-4bc6-4750-a532-f71f09933d71-kube-api-access-tqbgf\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548386 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-sockets\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548407 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548429 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548545 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-conf\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548636 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548770 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-startup\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548822 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-reloader\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.548875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmkqx\" (UniqueName: \"kubernetes.io/projected/98ad93ad-5869-4d8d-8074-c39600f257b1-kube-api-access-rmkqx\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.570084 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-tv4tz"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.574849 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.581939 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.585270 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-tv4tz"] Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.649897 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metrics-certs\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650197 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-conf\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650462 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650573 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-startup\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650764 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-reloader\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.650893 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmkqx\" (UniqueName: \"kubernetes.io/projected/98ad93ad-5869-4d8d-8074-c39600f257b1-kube-api-access-rmkqx\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqbgf\" (UniqueName: \"kubernetes.io/projected/555989d8-4bc6-4750-a532-f71f09933d71-kube-api-access-tqbgf\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651131 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metallb-excludel2\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651248 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-sockets\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651350 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651447 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f297\" (UniqueName: \"kubernetes.io/projected/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-kube-api-access-8f297\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.651547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.651744 4726 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.651861 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert podName:555989d8-4bc6-4750-a532-f71f09933d71 nodeName:}" failed. No retries permitted until 2025-10-04 03:55:49.151842085 +0000 UTC m=+923.326465288 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert") pod "frr-k8s-webhook-server-64bf5d555-76nkf" (UID: "555989d8-4bc6-4750-a532-f71f09933d71") : secret "frr-k8s-webhook-server-cert" not found Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.655392 4726 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.655598 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs podName:98ad93ad-5869-4d8d-8074-c39600f257b1 nodeName:}" failed. No retries permitted until 2025-10-04 03:55:49.155580764 +0000 UTC m=+923.330204097 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs") pod "frr-k8s-zktqf" (UID: "98ad93ad-5869-4d8d-8074-c39600f257b1") : secret "frr-k8s-certs-secret" not found Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.662384 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-conf\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.665589 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.665964 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-reloader\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.668516 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-startup\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.671535 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/98ad93ad-5869-4d8d-8074-c39600f257b1-frr-sockets\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.685425 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmkqx\" (UniqueName: \"kubernetes.io/projected/98ad93ad-5869-4d8d-8074-c39600f257b1-kube-api-access-rmkqx\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.685544 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqbgf\" (UniqueName: \"kubernetes.io/projected/555989d8-4bc6-4750-a532-f71f09933d71-kube-api-access-tqbgf\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.752822 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-cert\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.752882 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metallb-excludel2\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.752934 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f297\" (UniqueName: \"kubernetes.io/projected/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-kube-api-access-8f297\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.752980 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metrics-certs\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.753009 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtmrm\" (UniqueName: \"kubernetes.io/projected/eca76b37-13e9-4101-8486-7d009410a7bb-kube-api-access-jtmrm\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.753030 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-metrics-certs\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.753058 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.753209 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:55:48 crc kubenswrapper[4726]: E1004 03:55:48.753263 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist podName:ca2e3d23-2c1d-4b4a-85d6-c280ce107167 nodeName:}" failed. No retries permitted until 2025-10-04 03:55:49.253243995 +0000 UTC m=+923.427867208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist") pod "speaker-89wr2" (UID: "ca2e3d23-2c1d-4b4a-85d6-c280ce107167") : secret "metallb-memberlist" not found Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.753727 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metallb-excludel2\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.757166 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-metrics-certs\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.768785 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f297\" (UniqueName: \"kubernetes.io/projected/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-kube-api-access-8f297\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.854030 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtmrm\" (UniqueName: \"kubernetes.io/projected/eca76b37-13e9-4101-8486-7d009410a7bb-kube-api-access-jtmrm\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.854440 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-metrics-certs\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.854722 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-cert\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.858505 4726 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.859060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-metrics-certs\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.867875 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eca76b37-13e9-4101-8486-7d009410a7bb-cert\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.868794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtmrm\" (UniqueName: \"kubernetes.io/projected/eca76b37-13e9-4101-8486-7d009410a7bb-kube-api-access-jtmrm\") pod \"controller-68d546b9d8-tv4tz\" (UID: \"eca76b37-13e9-4101-8486-7d009410a7bb\") " pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:48 crc kubenswrapper[4726]: I1004 03:55:48.900920 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.083852 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-tv4tz"] Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.158227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.158290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.162581 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/98ad93ad-5869-4d8d-8074-c39600f257b1-metrics-certs\") pod \"frr-k8s-zktqf\" (UID: \"98ad93ad-5869-4d8d-8074-c39600f257b1\") " pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.162739 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/555989d8-4bc6-4750-a532-f71f09933d71-cert\") pod \"frr-k8s-webhook-server-64bf5d555-76nkf\" (UID: \"555989d8-4bc6-4750-a532-f71f09933d71\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.259224 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:49 crc kubenswrapper[4726]: E1004 03:55:49.259782 4726 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 03:55:49 crc kubenswrapper[4726]: E1004 03:55:49.259852 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist podName:ca2e3d23-2c1d-4b4a-85d6-c280ce107167 nodeName:}" failed. No retries permitted until 2025-10-04 03:55:50.259837492 +0000 UTC m=+924.434460705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist") pod "speaker-89wr2" (UID: "ca2e3d23-2c1d-4b4a-85d6-c280ce107167") : secret "metallb-memberlist" not found Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.386375 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.401690 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.736921 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf"] Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.893382 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" event={"ID":"555989d8-4bc6-4750-a532-f71f09933d71","Type":"ContainerStarted","Data":"d52bac908b3d9dabf468795d9d9d44a90ada9b816e3e245e0526020a906a4a0c"} Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.894678 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"5bbfd0c8a35b77fbe1b65c113c946162462ea25d0efc9c6ac4e45b0992fbf9d4"} Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.896686 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tv4tz" event={"ID":"eca76b37-13e9-4101-8486-7d009410a7bb","Type":"ContainerStarted","Data":"9d6c4806f6fb29d1695c4c290e361bb44287918ca0b8d88e74e51e89af074854"} Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.896751 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tv4tz" event={"ID":"eca76b37-13e9-4101-8486-7d009410a7bb","Type":"ContainerStarted","Data":"2f8a8d17837ad46dfe4021061f61ac3394c4a1edfa89f7756b0f07fdc3abb264"} Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.896778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-tv4tz" event={"ID":"eca76b37-13e9-4101-8486-7d009410a7bb","Type":"ContainerStarted","Data":"dc01f1d5f043d47110cb24941fad8591124541d8441e2e70237eea3fe5cecad5"} Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.897061 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:55:49 crc kubenswrapper[4726]: I1004 03:55:49.924212 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-tv4tz" podStartSLOduration=1.924193928 podStartE2EDuration="1.924193928s" podCreationTimestamp="2025-10-04 03:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:55:49.921274775 +0000 UTC m=+924.095897998" watchObservedRunningTime="2025-10-04 03:55:49.924193928 +0000 UTC m=+924.098817151" Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.276376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.281182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ca2e3d23-2c1d-4b4a-85d6-c280ce107167-memberlist\") pod \"speaker-89wr2\" (UID: \"ca2e3d23-2c1d-4b4a-85d6-c280ce107167\") " pod="metallb-system/speaker-89wr2" Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.355555 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-89wr2" Oct 04 03:55:50 crc kubenswrapper[4726]: W1004 03:55:50.373699 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca2e3d23_2c1d_4b4a_85d6_c280ce107167.slice/crio-eddcb3d1725f893389ccd64edee64aaddc3743a27d88577c0fb8b85cafe06f6f WatchSource:0}: Error finding container eddcb3d1725f893389ccd64edee64aaddc3743a27d88577c0fb8b85cafe06f6f: Status 404 returned error can't find the container with id eddcb3d1725f893389ccd64edee64aaddc3743a27d88577c0fb8b85cafe06f6f Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.907256 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89wr2" event={"ID":"ca2e3d23-2c1d-4b4a-85d6-c280ce107167","Type":"ContainerStarted","Data":"8fb01077e222d82cf1dcd03c4d07a0c56ff32ee66e20dc2223b6e4d74cb328d2"} Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.907606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89wr2" event={"ID":"ca2e3d23-2c1d-4b4a-85d6-c280ce107167","Type":"ContainerStarted","Data":"d11522f03def62e1c2a68a973967ef902617e799995723a813c63778d2dd2da8"} Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.907630 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-89wr2" event={"ID":"ca2e3d23-2c1d-4b4a-85d6-c280ce107167","Type":"ContainerStarted","Data":"eddcb3d1725f893389ccd64edee64aaddc3743a27d88577c0fb8b85cafe06f6f"} Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.908502 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-89wr2" Oct 04 03:55:50 crc kubenswrapper[4726]: I1004 03:55:50.928356 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-89wr2" podStartSLOduration=2.9283312329999998 podStartE2EDuration="2.928331233s" podCreationTimestamp="2025-10-04 03:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:55:50.925460192 +0000 UTC m=+925.100083415" watchObservedRunningTime="2025-10-04 03:55:50.928331233 +0000 UTC m=+925.102954456" Oct 04 03:55:56 crc kubenswrapper[4726]: I1004 03:55:56.955759 4726 generic.go:334] "Generic (PLEG): container finished" podID="98ad93ad-5869-4d8d-8074-c39600f257b1" containerID="0a29d12050c2c9ee4ba2bd86ceca8f46b5bc7c18c4a29c1cd2d50f1b0de57f8a" exitCode=0 Oct 04 03:55:56 crc kubenswrapper[4726]: I1004 03:55:56.955939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerDied","Data":"0a29d12050c2c9ee4ba2bd86ceca8f46b5bc7c18c4a29c1cd2d50f1b0de57f8a"} Oct 04 03:55:56 crc kubenswrapper[4726]: I1004 03:55:56.959582 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" event={"ID":"555989d8-4bc6-4750-a532-f71f09933d71","Type":"ContainerStarted","Data":"7aeb48323a1791ad02c244f013cea2dfa47bd38feb31bdf5122c5a260130f91f"} Oct 04 03:55:56 crc kubenswrapper[4726]: I1004 03:55:56.959821 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:55:57 crc kubenswrapper[4726]: I1004 03:55:57.027357 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" podStartSLOduration=2.628483142 podStartE2EDuration="9.027339283s" podCreationTimestamp="2025-10-04 03:55:48 +0000 UTC" firstStartedPulling="2025-10-04 03:55:49.744571704 +0000 UTC m=+923.919194917" lastFinishedPulling="2025-10-04 03:55:56.143427845 +0000 UTC m=+930.318051058" observedRunningTime="2025-10-04 03:55:57.026498226 +0000 UTC m=+931.201121449" watchObservedRunningTime="2025-10-04 03:55:57.027339283 +0000 UTC m=+931.201962496" Oct 04 03:55:57 crc kubenswrapper[4726]: I1004 03:55:57.967629 4726 generic.go:334] "Generic (PLEG): container finished" podID="98ad93ad-5869-4d8d-8074-c39600f257b1" containerID="2de94cf0c029f4fa0dc1fc32beeb74b44ac68789d5ecdf682fe81c236ed2811f" exitCode=0 Oct 04 03:55:57 crc kubenswrapper[4726]: I1004 03:55:57.967694 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerDied","Data":"2de94cf0c029f4fa0dc1fc32beeb74b44ac68789d5ecdf682fe81c236ed2811f"} Oct 04 03:55:58 crc kubenswrapper[4726]: I1004 03:55:58.974513 4726 generic.go:334] "Generic (PLEG): container finished" podID="98ad93ad-5869-4d8d-8074-c39600f257b1" containerID="1adabc229bfd0f16638afec66a4b6a3692a340aec67056fee8f0102abf179089" exitCode=0 Oct 04 03:55:58 crc kubenswrapper[4726]: I1004 03:55:58.974613 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerDied","Data":"1adabc229bfd0f16638afec66a4b6a3692a340aec67056fee8f0102abf179089"} Oct 04 03:55:59 crc kubenswrapper[4726]: I1004 03:55:59.993849 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"c1ebf2d86f3f2f2aa2d14183ca79f1b8872233baa1c112d51c6bcff360875243"} Oct 04 03:55:59 crc kubenswrapper[4726]: I1004 03:55:59.994236 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"14b3af871128de29598ba8d6a0084d71c473bc38729ec3c75ab93ab1ddeb24dc"} Oct 04 03:55:59 crc kubenswrapper[4726]: I1004 03:55:59.994252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"53ae276fa895e1202f625d841514d24058bdf47e83ab13e7d274a2a5ca3ef44c"} Oct 04 03:55:59 crc kubenswrapper[4726]: I1004 03:55:59.994267 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"405bfcc4b0ec0a98b46a418063169f5d0f7fb27ab31d09022f404dcb32781d8c"} Oct 04 03:55:59 crc kubenswrapper[4726]: I1004 03:55:59.994281 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"d0bd71b5885920b2e742c20fb7afab8b447f8d6b50c8c2fd8e7ede979bbed409"} Oct 04 03:56:00 crc kubenswrapper[4726]: I1004 03:56:00.359544 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-89wr2" Oct 04 03:56:01 crc kubenswrapper[4726]: I1004 03:56:01.024608 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zktqf" event={"ID":"98ad93ad-5869-4d8d-8074-c39600f257b1","Type":"ContainerStarted","Data":"6ebd32666a9af04236fb2d46b7b2f4886c9d3cb1908274863c2e0c6f2bdf80f5"} Oct 04 03:56:01 crc kubenswrapper[4726]: I1004 03:56:01.025958 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:56:01 crc kubenswrapper[4726]: I1004 03:56:01.059868 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zktqf" podStartSLOduration=6.521576416 podStartE2EDuration="13.059841062s" podCreationTimestamp="2025-10-04 03:55:48 +0000 UTC" firstStartedPulling="2025-10-04 03:55:49.568569737 +0000 UTC m=+923.743192970" lastFinishedPulling="2025-10-04 03:55:56.106834403 +0000 UTC m=+930.281457616" observedRunningTime="2025-10-04 03:56:01.058746167 +0000 UTC m=+935.233369410" watchObservedRunningTime="2025-10-04 03:56:01.059841062 +0000 UTC m=+935.234464315" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.477966 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.481581 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.484610 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bhstr" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.484792 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.484610 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.498000 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.657615 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j82b2\" (UniqueName: \"kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2\") pod \"openstack-operator-index-2xp5w\" (UID: \"99c0d99b-9a67-4019-a683-ed047d4c4f3b\") " pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.759385 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j82b2\" (UniqueName: \"kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2\") pod \"openstack-operator-index-2xp5w\" (UID: \"99c0d99b-9a67-4019-a683-ed047d4c4f3b\") " pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.782008 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j82b2\" (UniqueName: \"kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2\") pod \"openstack-operator-index-2xp5w\" (UID: \"99c0d99b-9a67-4019-a683-ed047d4c4f3b\") " pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:03 crc kubenswrapper[4726]: I1004 03:56:03.813143 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:04 crc kubenswrapper[4726]: I1004 03:56:04.048068 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:04 crc kubenswrapper[4726]: W1004 03:56:04.060361 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99c0d99b_9a67_4019_a683_ed047d4c4f3b.slice/crio-63e722e69eefef62de990125605d388d6487ebcc8a79765930b8cc60d62c2760 WatchSource:0}: Error finding container 63e722e69eefef62de990125605d388d6487ebcc8a79765930b8cc60d62c2760: Status 404 returned error can't find the container with id 63e722e69eefef62de990125605d388d6487ebcc8a79765930b8cc60d62c2760 Oct 04 03:56:04 crc kubenswrapper[4726]: I1004 03:56:04.187812 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:56:04 crc kubenswrapper[4726]: I1004 03:56:04.187872 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:56:04 crc kubenswrapper[4726]: I1004 03:56:04.387374 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:56:04 crc kubenswrapper[4726]: I1004 03:56:04.443781 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:56:05 crc kubenswrapper[4726]: I1004 03:56:05.050031 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xp5w" event={"ID":"99c0d99b-9a67-4019-a683-ed047d4c4f3b","Type":"ContainerStarted","Data":"63e722e69eefef62de990125605d388d6487ebcc8a79765930b8cc60d62c2760"} Oct 04 03:56:06 crc kubenswrapper[4726]: I1004 03:56:06.850646 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.065836 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xp5w" event={"ID":"99c0d99b-9a67-4019-a683-ed047d4c4f3b","Type":"ContainerStarted","Data":"ce900f9e8956f04a67614e41e27f39f4502a5f2958072efa546a7f96587ba3d4"} Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.098131 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2xp5w" podStartSLOduration=2.024734804 podStartE2EDuration="4.098080882s" podCreationTimestamp="2025-10-04 03:56:03 +0000 UTC" firstStartedPulling="2025-10-04 03:56:04.064920586 +0000 UTC m=+938.239543799" lastFinishedPulling="2025-10-04 03:56:06.138266664 +0000 UTC m=+940.312889877" observedRunningTime="2025-10-04 03:56:07.08918875 +0000 UTC m=+941.263812023" watchObservedRunningTime="2025-10-04 03:56:07.098080882 +0000 UTC m=+941.272704125" Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.463677 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fzc75"] Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.464941 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.483804 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fzc75"] Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.612517 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m59pp\" (UniqueName: \"kubernetes.io/projected/40e39198-082e-4c71-b67c-6d845e45f782-kube-api-access-m59pp\") pod \"openstack-operator-index-fzc75\" (UID: \"40e39198-082e-4c71-b67c-6d845e45f782\") " pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.713890 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m59pp\" (UniqueName: \"kubernetes.io/projected/40e39198-082e-4c71-b67c-6d845e45f782-kube-api-access-m59pp\") pod \"openstack-operator-index-fzc75\" (UID: \"40e39198-082e-4c71-b67c-6d845e45f782\") " pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.744527 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m59pp\" (UniqueName: \"kubernetes.io/projected/40e39198-082e-4c71-b67c-6d845e45f782-kube-api-access-m59pp\") pod \"openstack-operator-index-fzc75\" (UID: \"40e39198-082e-4c71-b67c-6d845e45f782\") " pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:07 crc kubenswrapper[4726]: I1004 03:56:07.803645 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:08 crc kubenswrapper[4726]: I1004 03:56:08.075721 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2xp5w" podUID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" containerName="registry-server" containerID="cri-o://ce900f9e8956f04a67614e41e27f39f4502a5f2958072efa546a7f96587ba3d4" gracePeriod=2 Oct 04 03:56:08 crc kubenswrapper[4726]: I1004 03:56:08.087016 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fzc75"] Oct 04 03:56:08 crc kubenswrapper[4726]: W1004 03:56:08.095959 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40e39198_082e_4c71_b67c_6d845e45f782.slice/crio-1e64a4293796a59d24518768ce626a5e8f457a8e53b099dc6e5a28112b608788 WatchSource:0}: Error finding container 1e64a4293796a59d24518768ce626a5e8f457a8e53b099dc6e5a28112b608788: Status 404 returned error can't find the container with id 1e64a4293796a59d24518768ce626a5e8f457a8e53b099dc6e5a28112b608788 Oct 04 03:56:08 crc kubenswrapper[4726]: I1004 03:56:08.907926 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-tv4tz" Oct 04 03:56:09 crc kubenswrapper[4726]: I1004 03:56:09.087884 4726 generic.go:334] "Generic (PLEG): container finished" podID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" containerID="ce900f9e8956f04a67614e41e27f39f4502a5f2958072efa546a7f96587ba3d4" exitCode=0 Oct 04 03:56:09 crc kubenswrapper[4726]: I1004 03:56:09.087979 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xp5w" event={"ID":"99c0d99b-9a67-4019-a683-ed047d4c4f3b","Type":"ContainerDied","Data":"ce900f9e8956f04a67614e41e27f39f4502a5f2958072efa546a7f96587ba3d4"} Oct 04 03:56:09 crc kubenswrapper[4726]: I1004 03:56:09.090148 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fzc75" event={"ID":"40e39198-082e-4c71-b67c-6d845e45f782","Type":"ContainerStarted","Data":"1e64a4293796a59d24518768ce626a5e8f457a8e53b099dc6e5a28112b608788"} Oct 04 03:56:09 crc kubenswrapper[4726]: I1004 03:56:09.392406 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zktqf" Oct 04 03:56:09 crc kubenswrapper[4726]: I1004 03:56:09.411013 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-76nkf" Oct 04 03:56:10 crc kubenswrapper[4726]: I1004 03:56:10.340897 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:10 crc kubenswrapper[4726]: I1004 03:56:10.451500 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j82b2\" (UniqueName: \"kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2\") pod \"99c0d99b-9a67-4019-a683-ed047d4c4f3b\" (UID: \"99c0d99b-9a67-4019-a683-ed047d4c4f3b\") " Oct 04 03:56:10 crc kubenswrapper[4726]: I1004 03:56:10.460304 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2" (OuterVolumeSpecName: "kube-api-access-j82b2") pod "99c0d99b-9a67-4019-a683-ed047d4c4f3b" (UID: "99c0d99b-9a67-4019-a683-ed047d4c4f3b"). InnerVolumeSpecName "kube-api-access-j82b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:56:10 crc kubenswrapper[4726]: I1004 03:56:10.553423 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j82b2\" (UniqueName: \"kubernetes.io/projected/99c0d99b-9a67-4019-a683-ed047d4c4f3b-kube-api-access-j82b2\") on node \"crc\" DevicePath \"\"" Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.108741 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fzc75" event={"ID":"40e39198-082e-4c71-b67c-6d845e45f782","Type":"ContainerStarted","Data":"fc20734b97344be20cbf811f398b8fc4b7861b6ac7464a03c43292dd53d7f646"} Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.112754 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xp5w" event={"ID":"99c0d99b-9a67-4019-a683-ed047d4c4f3b","Type":"ContainerDied","Data":"63e722e69eefef62de990125605d388d6487ebcc8a79765930b8cc60d62c2760"} Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.112826 4726 scope.go:117] "RemoveContainer" containerID="ce900f9e8956f04a67614e41e27f39f4502a5f2958072efa546a7f96587ba3d4" Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.112847 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xp5w" Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.150520 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fzc75" podStartSLOduration=2.197508568 podStartE2EDuration="4.150492054s" podCreationTimestamp="2025-10-04 03:56:07 +0000 UTC" firstStartedPulling="2025-10-04 03:56:08.103646073 +0000 UTC m=+942.278269296" lastFinishedPulling="2025-10-04 03:56:10.056629529 +0000 UTC m=+944.231252782" observedRunningTime="2025-10-04 03:56:11.134077833 +0000 UTC m=+945.308701086" watchObservedRunningTime="2025-10-04 03:56:11.150492054 +0000 UTC m=+945.325115297" Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.158982 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:11 crc kubenswrapper[4726]: I1004 03:56:11.167759 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2xp5w"] Oct 04 03:56:12 crc kubenswrapper[4726]: I1004 03:56:12.514874 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" path="/var/lib/kubelet/pods/99c0d99b-9a67-4019-a683-ed047d4c4f3b/volumes" Oct 04 03:56:17 crc kubenswrapper[4726]: I1004 03:56:17.804951 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:17 crc kubenswrapper[4726]: I1004 03:56:17.805442 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:17 crc kubenswrapper[4726]: I1004 03:56:17.852898 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:18 crc kubenswrapper[4726]: I1004 03:56:18.201835 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fzc75" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.496663 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x"] Oct 04 03:56:19 crc kubenswrapper[4726]: E1004 03:56:19.496942 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" containerName="registry-server" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.496958 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" containerName="registry-server" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.497178 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c0d99b-9a67-4019-a683-ed047d4c4f3b" containerName="registry-server" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.497956 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.499970 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-h5mz7" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.505579 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x"] Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.580854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.580914 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.580966 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvkxw\" (UniqueName: \"kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.681842 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvkxw\" (UniqueName: \"kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.681904 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.681939 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.682332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.682380 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.701723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvkxw\" (UniqueName: \"kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw\") pod \"1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:19 crc kubenswrapper[4726]: I1004 03:56:19.856428 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:20 crc kubenswrapper[4726]: I1004 03:56:20.336739 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x"] Oct 04 03:56:20 crc kubenswrapper[4726]: W1004 03:56:20.347981 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12b6c390_7847_4e46_bdd6_6f96a1780f18.slice/crio-a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3 WatchSource:0}: Error finding container a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3: Status 404 returned error can't find the container with id a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3 Oct 04 03:56:21 crc kubenswrapper[4726]: I1004 03:56:21.193605 4726 generic.go:334] "Generic (PLEG): container finished" podID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerID="7695fb91070e9c3a7a454bb68e80c1a4a1cfc0a0a4ddf5ed602ca30bd70efee4" exitCode=0 Oct 04 03:56:21 crc kubenswrapper[4726]: I1004 03:56:21.193668 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" event={"ID":"12b6c390-7847-4e46-bdd6-6f96a1780f18","Type":"ContainerDied","Data":"7695fb91070e9c3a7a454bb68e80c1a4a1cfc0a0a4ddf5ed602ca30bd70efee4"} Oct 04 03:56:21 crc kubenswrapper[4726]: I1004 03:56:21.193708 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" event={"ID":"12b6c390-7847-4e46-bdd6-6f96a1780f18","Type":"ContainerStarted","Data":"a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3"} Oct 04 03:56:22 crc kubenswrapper[4726]: I1004 03:56:22.203014 4726 generic.go:334] "Generic (PLEG): container finished" podID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerID="511292b17161a2ab1cbb357bc04b9f648446a86cf5997e107fcc909ee90d844f" exitCode=0 Oct 04 03:56:22 crc kubenswrapper[4726]: I1004 03:56:22.203064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" event={"ID":"12b6c390-7847-4e46-bdd6-6f96a1780f18","Type":"ContainerDied","Data":"511292b17161a2ab1cbb357bc04b9f648446a86cf5997e107fcc909ee90d844f"} Oct 04 03:56:23 crc kubenswrapper[4726]: I1004 03:56:23.214525 4726 generic.go:334] "Generic (PLEG): container finished" podID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerID="bebe895103fbf19572c33a9b52c1fa0d14cc92e1dba072cecdba283cb4707819" exitCode=0 Oct 04 03:56:23 crc kubenswrapper[4726]: I1004 03:56:23.214636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" event={"ID":"12b6c390-7847-4e46-bdd6-6f96a1780f18","Type":"ContainerDied","Data":"bebe895103fbf19572c33a9b52c1fa0d14cc92e1dba072cecdba283cb4707819"} Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.542010 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.675677 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvkxw\" (UniqueName: \"kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw\") pod \"12b6c390-7847-4e46-bdd6-6f96a1780f18\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.676176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util\") pod \"12b6c390-7847-4e46-bdd6-6f96a1780f18\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.676251 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle\") pod \"12b6c390-7847-4e46-bdd6-6f96a1780f18\" (UID: \"12b6c390-7847-4e46-bdd6-6f96a1780f18\") " Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.676949 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle" (OuterVolumeSpecName: "bundle") pod "12b6c390-7847-4e46-bdd6-6f96a1780f18" (UID: "12b6c390-7847-4e46-bdd6-6f96a1780f18"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.693142 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util" (OuterVolumeSpecName: "util") pod "12b6c390-7847-4e46-bdd6-6f96a1780f18" (UID: "12b6c390-7847-4e46-bdd6-6f96a1780f18"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.693258 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw" (OuterVolumeSpecName: "kube-api-access-zvkxw") pod "12b6c390-7847-4e46-bdd6-6f96a1780f18" (UID: "12b6c390-7847-4e46-bdd6-6f96a1780f18"). InnerVolumeSpecName "kube-api-access-zvkxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.778172 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvkxw\" (UniqueName: \"kubernetes.io/projected/12b6c390-7847-4e46-bdd6-6f96a1780f18-kube-api-access-zvkxw\") on node \"crc\" DevicePath \"\"" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.778240 4726 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-util\") on node \"crc\" DevicePath \"\"" Oct 04 03:56:24 crc kubenswrapper[4726]: I1004 03:56:24.778267 4726 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/12b6c390-7847-4e46-bdd6-6f96a1780f18-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:56:25 crc kubenswrapper[4726]: I1004 03:56:25.240312 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" event={"ID":"12b6c390-7847-4e46-bdd6-6f96a1780f18","Type":"ContainerDied","Data":"a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3"} Oct 04 03:56:25 crc kubenswrapper[4726]: I1004 03:56:25.240383 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x" Oct 04 03:56:25 crc kubenswrapper[4726]: I1004 03:56:25.240958 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2d6c39fb9315e768a22360535129349151c8b81efb553b4bd3b5a5125a92de3" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.180828 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx"] Oct 04 03:56:32 crc kubenswrapper[4726]: E1004 03:56:32.182476 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="pull" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.182543 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="pull" Oct 04 03:56:32 crc kubenswrapper[4726]: E1004 03:56:32.182617 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="util" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.182669 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="util" Oct 04 03:56:32 crc kubenswrapper[4726]: E1004 03:56:32.182722 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="extract" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.182781 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="extract" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.182970 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="12b6c390-7847-4e46-bdd6-6f96a1780f18" containerName="extract" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.183652 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.185640 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-sd49t" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.205748 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx"] Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.312814 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5hjl\" (UniqueName: \"kubernetes.io/projected/0bd87ab8-94f6-41a4-b76c-7774cac7b625-kube-api-access-d5hjl\") pod \"openstack-operator-controller-operator-5499555f6-blzgx\" (UID: \"0bd87ab8-94f6-41a4-b76c-7774cac7b625\") " pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.414205 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5hjl\" (UniqueName: \"kubernetes.io/projected/0bd87ab8-94f6-41a4-b76c-7774cac7b625-kube-api-access-d5hjl\") pod \"openstack-operator-controller-operator-5499555f6-blzgx\" (UID: \"0bd87ab8-94f6-41a4-b76c-7774cac7b625\") " pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.440023 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5hjl\" (UniqueName: \"kubernetes.io/projected/0bd87ab8-94f6-41a4-b76c-7774cac7b625-kube-api-access-d5hjl\") pod \"openstack-operator-controller-operator-5499555f6-blzgx\" (UID: \"0bd87ab8-94f6-41a4-b76c-7774cac7b625\") " pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.500326 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:32 crc kubenswrapper[4726]: I1004 03:56:32.763573 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx"] Oct 04 03:56:33 crc kubenswrapper[4726]: I1004 03:56:33.306284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" event={"ID":"0bd87ab8-94f6-41a4-b76c-7774cac7b625","Type":"ContainerStarted","Data":"67bf7d0c2f04e9339e4a2411c8341998a128fa8418ccb4e9c4763503e14ac520"} Oct 04 03:56:34 crc kubenswrapper[4726]: I1004 03:56:34.188740 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:56:34 crc kubenswrapper[4726]: I1004 03:56:34.188833 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:56:34 crc kubenswrapper[4726]: I1004 03:56:34.188908 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 03:56:34 crc kubenswrapper[4726]: I1004 03:56:34.189869 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 03:56:34 crc kubenswrapper[4726]: I1004 03:56:34.190000 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7" gracePeriod=600 Oct 04 03:56:35 crc kubenswrapper[4726]: I1004 03:56:35.321049 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7" exitCode=0 Oct 04 03:56:35 crc kubenswrapper[4726]: I1004 03:56:35.321135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7"} Oct 04 03:56:35 crc kubenswrapper[4726]: I1004 03:56:35.321172 4726 scope.go:117] "RemoveContainer" containerID="1de7b1c78dbd159a44c6e91af95939dea88d8fd160c5e27b5f942aa7de5fa5db" Oct 04 03:56:37 crc kubenswrapper[4726]: I1004 03:56:37.341755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea"} Oct 04 03:56:37 crc kubenswrapper[4726]: I1004 03:56:37.344565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" event={"ID":"0bd87ab8-94f6-41a4-b76c-7774cac7b625","Type":"ContainerStarted","Data":"8388882be3f72cd35d4dcf230018d6977b41be16900488190f9b75f854a1184c"} Oct 04 03:56:39 crc kubenswrapper[4726]: I1004 03:56:39.357098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" event={"ID":"0bd87ab8-94f6-41a4-b76c-7774cac7b625","Type":"ContainerStarted","Data":"cc82a26010775f0508f4f37007863f770ef39315dd2244a214238bec857a35b6"} Oct 04 03:56:39 crc kubenswrapper[4726]: I1004 03:56:39.357790 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:39 crc kubenswrapper[4726]: I1004 03:56:39.405758 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" podStartSLOduration=1.670917635 podStartE2EDuration="7.405733654s" podCreationTimestamp="2025-10-04 03:56:32 +0000 UTC" firstStartedPulling="2025-10-04 03:56:32.774740132 +0000 UTC m=+966.949363365" lastFinishedPulling="2025-10-04 03:56:38.509556171 +0000 UTC m=+972.684179384" observedRunningTime="2025-10-04 03:56:39.401046256 +0000 UTC m=+973.575669509" watchObservedRunningTime="2025-10-04 03:56:39.405733654 +0000 UTC m=+973.580356897" Oct 04 03:56:42 crc kubenswrapper[4726]: I1004 03:56:42.525252 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5499555f6-blzgx" Oct 04 03:56:59 crc kubenswrapper[4726]: I1004 03:56:59.984278 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5"] Oct 04 03:56:59 crc kubenswrapper[4726]: I1004 03:56:59.985959 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:56:59 crc kubenswrapper[4726]: I1004 03:56:59.990411 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-lg4w8" Oct 04 03:56:59 crc kubenswrapper[4726]: I1004 03:56:59.990547 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27"] Oct 04 03:56:59 crc kubenswrapper[4726]: I1004 03:56:59.994296 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.003073 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vmlrv" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.014440 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.057618 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.071038 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.071553 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d88cj\" (UniqueName: \"kubernetes.io/projected/f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3-kube-api-access-d88cj\") pod \"cinder-operator-controller-manager-7d4d4f8d-tzr27\" (UID: \"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.072798 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.098115 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.098215 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-85xpk" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.102055 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.103100 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.106200 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.123369 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.128960 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.130375 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.132768 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-54b26" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.135012 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.136048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.141283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-w4bjj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.145544 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dc4wt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.147011 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.150742 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.151644 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.156806 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.157168 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-smbgs" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.168354 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.175246 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d88cj\" (UniqueName: \"kubernetes.io/projected/f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3-kube-api-access-d88cj\") pod \"cinder-operator-controller-manager-7d4d4f8d-tzr27\" (UID: \"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.175608 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk224\" (UniqueName: \"kubernetes.io/projected/1e47df33-2621-4cfe-9c83-4d4bd1d616aa-kube-api-access-tk224\") pod \"barbican-operator-controller-manager-5f7c849b98-mfsl5\" (UID: \"1e47df33-2621-4cfe-9c83-4d4bd1d616aa\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.183427 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.184764 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.192743 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-6z7gv" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.197786 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.198766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.205720 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-5nkjn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.213940 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.224842 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d88cj\" (UniqueName: \"kubernetes.io/projected/f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3-kube-api-access-d88cj\") pod \"cinder-operator-controller-manager-7d4d4f8d-tzr27\" (UID: \"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.250031 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.257995 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.260281 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.264904 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mfx9m" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.270181 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276220 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-586bv\" (UniqueName: \"kubernetes.io/projected/5e5f2785-db94-4fdc-9cb2-10b1ce940694-kube-api-access-586bv\") pod \"horizon-operator-controller-manager-54876c876f-fvw5p\" (UID: \"5e5f2785-db94-4fdc-9cb2-10b1ce940694\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk224\" (UniqueName: \"kubernetes.io/projected/1e47df33-2621-4cfe-9c83-4d4bd1d616aa-kube-api-access-tk224\") pod \"barbican-operator-controller-manager-5f7c849b98-mfsl5\" (UID: \"1e47df33-2621-4cfe-9c83-4d4bd1d616aa\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276283 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsb5h\" (UniqueName: \"kubernetes.io/projected/79a719e7-5bce-424a-b837-ab0bca45936e-kube-api-access-fsb5h\") pod \"glance-operator-controller-manager-5568b5d68-cjx6d\" (UID: \"79a719e7-5bce-424a-b837-ab0bca45936e\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276337 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwwnj\" (UniqueName: \"kubernetes.io/projected/a520bf75-9223-4209-87d4-248ce3e0d190-kube-api-access-pwwnj\") pod \"designate-operator-controller-manager-75dfd9b554-gtshn\" (UID: \"a520bf75-9223-4209-87d4-248ce3e0d190\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276354 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276374 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmjpz\" (UniqueName: \"kubernetes.io/projected/ea261e29-40c5-4a26-a02f-8a09e0c7320d-kube-api-access-mmjpz\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.276392 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldld7\" (UniqueName: \"kubernetes.io/projected/ac49881b-2d45-4ba0-8b60-8461a2d5f35b-kube-api-access-ldld7\") pod \"heat-operator-controller-manager-8f58bc9db-lqs6l\" (UID: \"ac49881b-2d45-4ba0-8b60-8461a2d5f35b\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.285248 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.286308 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.292634 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4w8fh" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.304601 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.305598 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.317052 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.320674 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.321643 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-btkpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.322726 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk224\" (UniqueName: \"kubernetes.io/projected/1e47df33-2621-4cfe-9c83-4d4bd1d616aa-kube-api-access-tk224\") pod \"barbican-operator-controller-manager-5f7c849b98-mfsl5\" (UID: \"1e47df33-2621-4cfe-9c83-4d4bd1d616aa\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.334280 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.334909 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.336407 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.346024 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bbb7p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.353130 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.354777 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.358424 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-f2s4k" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.361741 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378449 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-586bv\" (UniqueName: \"kubernetes.io/projected/5e5f2785-db94-4fdc-9cb2-10b1ce940694-kube-api-access-586bv\") pod \"horizon-operator-controller-manager-54876c876f-fvw5p\" (UID: \"5e5f2785-db94-4fdc-9cb2-10b1ce940694\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378503 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jsn\" (UniqueName: \"kubernetes.io/projected/05b7da99-96e2-40be-8fc1-f847b8e7a643-kube-api-access-62jsn\") pod \"nova-operator-controller-manager-7c7fc454ff-mnndm\" (UID: \"05b7da99-96e2-40be-8fc1-f847b8e7a643\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378528 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsb5h\" (UniqueName: \"kubernetes.io/projected/79a719e7-5bce-424a-b837-ab0bca45936e-kube-api-access-fsb5h\") pod \"glance-operator-controller-manager-5568b5d68-cjx6d\" (UID: \"79a719e7-5bce-424a-b837-ab0bca45936e\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378549 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtdw6\" (UniqueName: \"kubernetes.io/projected/488fc85d-f45f-4d8d-93e2-b58ba9306a76-kube-api-access-gtdw6\") pod \"keystone-operator-controller-manager-655d88ccb9-wwdj2\" (UID: \"488fc85d-f45f-4d8d-93e2-b58ba9306a76\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378597 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jftw6\" (UniqueName: \"kubernetes.io/projected/d59f6f6b-419b-4ce3-adb6-11381aed6a5d-kube-api-access-jftw6\") pod \"neutron-operator-controller-manager-8d984cc4d-lqsxr\" (UID: \"d59f6f6b-419b-4ce3-adb6-11381aed6a5d\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpj2q\" (UniqueName: \"kubernetes.io/projected/cadfc4b1-9327-4a55-8544-44422572319c-kube-api-access-xpj2q\") pod \"octavia-operator-controller-manager-7468f855d8-zkgst\" (UID: \"cadfc4b1-9327-4a55-8544-44422572319c\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378662 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmv2t\" (UniqueName: \"kubernetes.io/projected/9f398c24-0435-49f5-915c-a54efeadea0c-kube-api-access-pmv2t\") pod \"ironic-operator-controller-manager-699b87f775-v72j6\" (UID: \"9f398c24-0435-49f5-915c-a54efeadea0c\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378686 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brvp4\" (UniqueName: \"kubernetes.io/projected/6508ae63-4f75-496f-b7b6-5dcd8a0cb008-kube-api-access-brvp4\") pod \"manila-operator-controller-manager-65d89cfd9f-6dpjp\" (UID: \"6508ae63-4f75-496f-b7b6-5dcd8a0cb008\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378735 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwwnj\" (UniqueName: \"kubernetes.io/projected/a520bf75-9223-4209-87d4-248ce3e0d190-kube-api-access-pwwnj\") pod \"designate-operator-controller-manager-75dfd9b554-gtshn\" (UID: \"a520bf75-9223-4209-87d4-248ce3e0d190\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378753 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378773 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmjpz\" (UniqueName: \"kubernetes.io/projected/ea261e29-40c5-4a26-a02f-8a09e0c7320d-kube-api-access-mmjpz\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378805 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhxx5\" (UniqueName: \"kubernetes.io/projected/d55e66e5-0c46-4df2-909c-fc739a26c5cc-kube-api-access-nhxx5\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-s67gj\" (UID: \"d55e66e5-0c46-4df2-909c-fc739a26c5cc\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.378826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldld7\" (UniqueName: \"kubernetes.io/projected/ac49881b-2d45-4ba0-8b60-8461a2d5f35b-kube-api-access-ldld7\") pod \"heat-operator-controller-manager-8f58bc9db-lqs6l\" (UID: \"ac49881b-2d45-4ba0-8b60-8461a2d5f35b\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.379275 4726 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.379322 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert podName:ea261e29-40c5-4a26-a02f-8a09e0c7320d nodeName:}" failed. No retries permitted until 2025-10-04 03:57:00.879306255 +0000 UTC m=+995.053929468 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert") pod "infra-operator-controller-manager-7ddb876846-kgv8l" (UID: "ea261e29-40c5-4a26-a02f-8a09e0c7320d") : secret "infra-operator-webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.403953 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.405259 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.414310 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.415199 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hqjzx" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.415884 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsb5h\" (UniqueName: \"kubernetes.io/projected/79a719e7-5bce-424a-b837-ab0bca45936e-kube-api-access-fsb5h\") pod \"glance-operator-controller-manager-5568b5d68-cjx6d\" (UID: \"79a719e7-5bce-424a-b837-ab0bca45936e\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.419626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwwnj\" (UniqueName: \"kubernetes.io/projected/a520bf75-9223-4209-87d4-248ce3e0d190-kube-api-access-pwwnj\") pod \"designate-operator-controller-manager-75dfd9b554-gtshn\" (UID: \"a520bf75-9223-4209-87d4-248ce3e0d190\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.423258 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmjpz\" (UniqueName: \"kubernetes.io/projected/ea261e29-40c5-4a26-a02f-8a09e0c7320d-kube-api-access-mmjpz\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.430590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-586bv\" (UniqueName: \"kubernetes.io/projected/5e5f2785-db94-4fdc-9cb2-10b1ce940694-kube-api-access-586bv\") pod \"horizon-operator-controller-manager-54876c876f-fvw5p\" (UID: \"5e5f2785-db94-4fdc-9cb2-10b1ce940694\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.430936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldld7\" (UniqueName: \"kubernetes.io/projected/ac49881b-2d45-4ba0-8b60-8461a2d5f35b-kube-api-access-ldld7\") pod \"heat-operator-controller-manager-8f58bc9db-lqs6l\" (UID: \"ac49881b-2d45-4ba0-8b60-8461a2d5f35b\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.439207 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.444698 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.449574 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.450571 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.454168 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2mttq" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.461977 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.466149 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483476 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhxx5\" (UniqueName: \"kubernetes.io/projected/d55e66e5-0c46-4df2-909c-fc739a26c5cc-kube-api-access-nhxx5\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-s67gj\" (UID: \"d55e66e5-0c46-4df2-909c-fc739a26c5cc\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483540 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jsn\" (UniqueName: \"kubernetes.io/projected/05b7da99-96e2-40be-8fc1-f847b8e7a643-kube-api-access-62jsn\") pod \"nova-operator-controller-manager-7c7fc454ff-mnndm\" (UID: \"05b7da99-96e2-40be-8fc1-f847b8e7a643\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483564 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtdw6\" (UniqueName: \"kubernetes.io/projected/488fc85d-f45f-4d8d-93e2-b58ba9306a76-kube-api-access-gtdw6\") pod \"keystone-operator-controller-manager-655d88ccb9-wwdj2\" (UID: \"488fc85d-f45f-4d8d-93e2-b58ba9306a76\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483583 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jftw6\" (UniqueName: \"kubernetes.io/projected/d59f6f6b-419b-4ce3-adb6-11381aed6a5d-kube-api-access-jftw6\") pod \"neutron-operator-controller-manager-8d984cc4d-lqsxr\" (UID: \"d59f6f6b-419b-4ce3-adb6-11381aed6a5d\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpj2q\" (UniqueName: \"kubernetes.io/projected/cadfc4b1-9327-4a55-8544-44422572319c-kube-api-access-xpj2q\") pod \"octavia-operator-controller-manager-7468f855d8-zkgst\" (UID: \"cadfc4b1-9327-4a55-8544-44422572319c\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483668 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmv2t\" (UniqueName: \"kubernetes.io/projected/9f398c24-0435-49f5-915c-a54efeadea0c-kube-api-access-pmv2t\") pod \"ironic-operator-controller-manager-699b87f775-v72j6\" (UID: \"9f398c24-0435-49f5-915c-a54efeadea0c\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.483698 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brvp4\" (UniqueName: \"kubernetes.io/projected/6508ae63-4f75-496f-b7b6-5dcd8a0cb008-kube-api-access-brvp4\") pod \"manila-operator-controller-manager-65d89cfd9f-6dpjp\" (UID: \"6508ae63-4f75-496f-b7b6-5dcd8a0cb008\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.487230 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.494024 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.537782 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jsn\" (UniqueName: \"kubernetes.io/projected/05b7da99-96e2-40be-8fc1-f847b8e7a643-kube-api-access-62jsn\") pod \"nova-operator-controller-manager-7c7fc454ff-mnndm\" (UID: \"05b7da99-96e2-40be-8fc1-f847b8e7a643\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.539300 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhxx5\" (UniqueName: \"kubernetes.io/projected/d55e66e5-0c46-4df2-909c-fc739a26c5cc-kube-api-access-nhxx5\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-s67gj\" (UID: \"d55e66e5-0c46-4df2-909c-fc739a26c5cc\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.546788 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtdw6\" (UniqueName: \"kubernetes.io/projected/488fc85d-f45f-4d8d-93e2-b58ba9306a76-kube-api-access-gtdw6\") pod \"keystone-operator-controller-manager-655d88ccb9-wwdj2\" (UID: \"488fc85d-f45f-4d8d-93e2-b58ba9306a76\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.547642 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jftw6\" (UniqueName: \"kubernetes.io/projected/d59f6f6b-419b-4ce3-adb6-11381aed6a5d-kube-api-access-jftw6\") pod \"neutron-operator-controller-manager-8d984cc4d-lqsxr\" (UID: \"d59f6f6b-419b-4ce3-adb6-11381aed6a5d\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.555050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmv2t\" (UniqueName: \"kubernetes.io/projected/9f398c24-0435-49f5-915c-a54efeadea0c-kube-api-access-pmv2t\") pod \"ironic-operator-controller-manager-699b87f775-v72j6\" (UID: \"9f398c24-0435-49f5-915c-a54efeadea0c\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.563053 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.564607 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpj2q\" (UniqueName: \"kubernetes.io/projected/cadfc4b1-9327-4a55-8544-44422572319c-kube-api-access-xpj2q\") pod \"octavia-operator-controller-manager-7468f855d8-zkgst\" (UID: \"cadfc4b1-9327-4a55-8544-44422572319c\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.568522 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brvp4\" (UniqueName: \"kubernetes.io/projected/6508ae63-4f75-496f-b7b6-5dcd8a0cb008-kube-api-access-brvp4\") pod \"manila-operator-controller-manager-65d89cfd9f-6dpjp\" (UID: \"6508ae63-4f75-496f-b7b6-5dcd8a0cb008\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.574836 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.582507 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.588671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz5j2\" (UniqueName: \"kubernetes.io/projected/6b245f13-4e4b-4a41-b148-482d581e2c5c-kube-api-access-fz5j2\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.588756 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.588934 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmhss\" (UniqueName: \"kubernetes.io/projected/e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828-kube-api-access-gmhss\") pod \"ovn-operator-controller-manager-579449c7d5-8pnpc\" (UID: \"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.618621 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.629960 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.631096 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.631377 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.641424 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-8nz2n" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.641833 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.642618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.642853 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.642953 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.645266 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6jkml" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.645400 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-tl9fn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.647060 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.652094 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.672516 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.683780 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.684794 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.691418 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-gqvv9" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.708724 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.709434 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmhss\" (UniqueName: \"kubernetes.io/projected/e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828-kube-api-access-gmhss\") pod \"ovn-operator-controller-manager-579449c7d5-8pnpc\" (UID: \"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.709730 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wssp\" (UniqueName: \"kubernetes.io/projected/efa34c5f-e531-4319-b5fb-7497c4d026f0-kube-api-access-6wssp\") pod \"placement-operator-controller-manager-54689d9f88-lsxtw\" (UID: \"efa34c5f-e531-4319-b5fb-7497c4d026f0\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.709839 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzd6r\" (UniqueName: \"kubernetes.io/projected/42949778-256b-4809-b2ce-d633428e649c-kube-api-access-dzd6r\") pod \"telemetry-operator-controller-manager-5d4d74dd89-2j9rb\" (UID: \"42949778-256b-4809-b2ce-d633428e649c\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.710492 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlsdb\" (UniqueName: \"kubernetes.io/projected/392e7503-f32e-4e8f-ac07-d75844253ea1-kube-api-access-vlsdb\") pod \"test-operator-controller-manager-5cd5cb47d7-4v7k8\" (UID: \"392e7503-f32e-4e8f-ac07-d75844253ea1\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.710591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz5j2\" (UniqueName: \"kubernetes.io/projected/6b245f13-4e4b-4a41-b148-482d581e2c5c-kube-api-access-fz5j2\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.710948 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvnvt\" (UniqueName: \"kubernetes.io/projected/5cb54ad2-85db-4291-87dc-4f0b9ea688c0-kube-api-access-vvnvt\") pod \"swift-operator-controller-manager-6859f9b676-hv9lz\" (UID: \"5cb54ad2-85db-4291-87dc-4f0b9ea688c0\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.711021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.711248 4726 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.711539 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert podName:6b245f13-4e4b-4a41-b148-482d581e2c5c nodeName:}" failed. No retries permitted until 2025-10-04 03:57:01.211520876 +0000 UTC m=+995.386144089 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" (UID: "6b245f13-4e4b-4a41-b148-482d581e2c5c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.715772 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.722073 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.732890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmhss\" (UniqueName: \"kubernetes.io/projected/e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828-kube-api-access-gmhss\") pod \"ovn-operator-controller-manager-579449c7d5-8pnpc\" (UID: \"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.741479 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.742545 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.750520 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lbnph" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.756644 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.760078 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz5j2\" (UniqueName: \"kubernetes.io/projected/6b245f13-4e4b-4a41-b148-482d581e2c5c-kube-api-access-fz5j2\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.776775 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.788612 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.789722 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.793386 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.793686 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gbzkm" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.793991 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.800408 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.817864 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5xzb\" (UniqueName: \"kubernetes.io/projected/262abd5f-25d1-488f-b864-f9cc308a68f4-kube-api-access-b5xzb\") pod \"watcher-operator-controller-manager-6cbc6dd547-kn2pc\" (UID: \"262abd5f-25d1-488f-b864-f9cc308a68f4\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.817939 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wssp\" (UniqueName: \"kubernetes.io/projected/efa34c5f-e531-4319-b5fb-7497c4d026f0-kube-api-access-6wssp\") pod \"placement-operator-controller-manager-54689d9f88-lsxtw\" (UID: \"efa34c5f-e531-4319-b5fb-7497c4d026f0\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.817986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzd6r\" (UniqueName: \"kubernetes.io/projected/42949778-256b-4809-b2ce-d633428e649c-kube-api-access-dzd6r\") pod \"telemetry-operator-controller-manager-5d4d74dd89-2j9rb\" (UID: \"42949778-256b-4809-b2ce-d633428e649c\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.818017 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlsdb\" (UniqueName: \"kubernetes.io/projected/392e7503-f32e-4e8f-ac07-d75844253ea1-kube-api-access-vlsdb\") pod \"test-operator-controller-manager-5cd5cb47d7-4v7k8\" (UID: \"392e7503-f32e-4e8f-ac07-d75844253ea1\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.818043 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5fmp\" (UniqueName: \"kubernetes.io/projected/aaa6fd43-d6ed-4d86-9827-27d2529499e3-kube-api-access-x5fmp\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.818068 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvnvt\" (UniqueName: \"kubernetes.io/projected/5cb54ad2-85db-4291-87dc-4f0b9ea688c0-kube-api-access-vvnvt\") pod \"swift-operator-controller-manager-6859f9b676-hv9lz\" (UID: \"5cb54ad2-85db-4291-87dc-4f0b9ea688c0\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.818103 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.837351 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.838459 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.843659 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-zn2px" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.857610 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wssp\" (UniqueName: \"kubernetes.io/projected/efa34c5f-e531-4319-b5fb-7497c4d026f0-kube-api-access-6wssp\") pod \"placement-operator-controller-manager-54689d9f88-lsxtw\" (UID: \"efa34c5f-e531-4319-b5fb-7497c4d026f0\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.859593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvnvt\" (UniqueName: \"kubernetes.io/projected/5cb54ad2-85db-4291-87dc-4f0b9ea688c0-kube-api-access-vvnvt\") pod \"swift-operator-controller-manager-6859f9b676-hv9lz\" (UID: \"5cb54ad2-85db-4291-87dc-4f0b9ea688c0\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.863736 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzd6r\" (UniqueName: \"kubernetes.io/projected/42949778-256b-4809-b2ce-d633428e649c-kube-api-access-dzd6r\") pod \"telemetry-operator-controller-manager-5d4d74dd89-2j9rb\" (UID: \"42949778-256b-4809-b2ce-d633428e649c\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.864015 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.873985 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlsdb\" (UniqueName: \"kubernetes.io/projected/392e7503-f32e-4e8f-ac07-d75844253ea1-kube-api-access-vlsdb\") pod \"test-operator-controller-manager-5cd5cb47d7-4v7k8\" (UID: \"392e7503-f32e-4e8f-ac07-d75844253ea1\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.877494 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7"] Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.920801 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5fmp\" (UniqueName: \"kubernetes.io/projected/aaa6fd43-d6ed-4d86-9827-27d2529499e3-kube-api-access-x5fmp\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.920867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.920912 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5xzb\" (UniqueName: \"kubernetes.io/projected/262abd5f-25d1-488f-b864-f9cc308a68f4-kube-api-access-b5xzb\") pod \"watcher-operator-controller-manager-6cbc6dd547-kn2pc\" (UID: \"262abd5f-25d1-488f-b864-f9cc308a68f4\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.920931 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.920967 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8lv6\" (UniqueName: \"kubernetes.io/projected/962ba8ee-fbcf-4ef5-b740-bd4a782d0881-kube-api-access-c8lv6\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7\" (UID: \"962ba8ee-fbcf-4ef5-b740-bd4a782d0881\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.922690 4726 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: E1004 03:57:00.922748 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert podName:aaa6fd43-d6ed-4d86-9827-27d2529499e3 nodeName:}" failed. No retries permitted until 2025-10-04 03:57:01.422731602 +0000 UTC m=+995.597354815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert") pod "openstack-operator-controller-manager-ff7c6f545-hgqmd" (UID: "aaa6fd43-d6ed-4d86-9827-27d2529499e3") : secret "webhook-server-cert" not found Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.933906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ea261e29-40c5-4a26-a02f-8a09e0c7320d-cert\") pod \"infra-operator-controller-manager-7ddb876846-kgv8l\" (UID: \"ea261e29-40c5-4a26-a02f-8a09e0c7320d\") " pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.947231 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5fmp\" (UniqueName: \"kubernetes.io/projected/aaa6fd43-d6ed-4d86-9827-27d2529499e3-kube-api-access-x5fmp\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.956977 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5xzb\" (UniqueName: \"kubernetes.io/projected/262abd5f-25d1-488f-b864-f9cc308a68f4-kube-api-access-b5xzb\") pod \"watcher-operator-controller-manager-6cbc6dd547-kn2pc\" (UID: \"262abd5f-25d1-488f-b864-f9cc308a68f4\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.981640 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:00 crc kubenswrapper[4726]: I1004 03:57:00.985597 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.020087 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.020544 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.026263 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8lv6\" (UniqueName: \"kubernetes.io/projected/962ba8ee-fbcf-4ef5-b740-bd4a782d0881-kube-api-access-c8lv6\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7\" (UID: \"962ba8ee-fbcf-4ef5-b740-bd4a782d0881\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.030449 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.035345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.053830 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8lv6\" (UniqueName: \"kubernetes.io/projected/962ba8ee-fbcf-4ef5-b740-bd4a782d0881-kube-api-access-c8lv6\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7\" (UID: \"962ba8ee-fbcf-4ef5-b740-bd4a782d0881\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" Oct 04 03:57:01 crc kubenswrapper[4726]: W1004 03:57:01.073084 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e5f2785_db94_4fdc_9cb2_10b1ce940694.slice/crio-b0942e954ef63b44f02c34c73814353a17a7532d12dedd22925b8157b525f5bb WatchSource:0}: Error finding container b0942e954ef63b44f02c34c73814353a17a7532d12dedd22925b8157b525f5bb: Status 404 returned error can't find the container with id b0942e954ef63b44f02c34c73814353a17a7532d12dedd22925b8157b525f5bb Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.094299 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.105352 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.111087 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.179427 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.230433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.236481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b245f13-4e4b-4a41-b148-482d581e2c5c-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt\" (UID: \"6b245f13-4e4b-4a41-b148-482d581e2c5c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.277966 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.440226 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.440459 4726 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.440549 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert podName:aaa6fd43-d6ed-4d86-9827-27d2529499e3 nodeName:}" failed. No retries permitted until 2025-10-04 03:57:02.440523483 +0000 UTC m=+996.615146696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert") pod "openstack-operator-controller-manager-ff7c6f545-hgqmd" (UID: "aaa6fd43-d6ed-4d86-9827-27d2529499e3") : secret "webhook-server-cert" not found Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.448409 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.490414 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.515414 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.531752 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.537705 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.564566 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.569703 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.572933 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" event={"ID":"5e5f2785-db94-4fdc-9cb2-10b1ce940694","Type":"ContainerStarted","Data":"b0942e954ef63b44f02c34c73814353a17a7532d12dedd22925b8157b525f5bb"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.575396 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" event={"ID":"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3","Type":"ContainerStarted","Data":"5abfb7ca723716b356505170e699c8db7aee3ca5e7830a4eea5e0ddc2e403f17"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.576656 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" event={"ID":"a520bf75-9223-4209-87d4-248ce3e0d190","Type":"ContainerStarted","Data":"feb9426158a62ebd501bcbb947e438073b3c5e40fbb6ab3dcce34d51288f65f9"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.579304 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" event={"ID":"488fc85d-f45f-4d8d-93e2-b58ba9306a76","Type":"ContainerStarted","Data":"7b6d89422c8bf98d6df0b3e07ae8037e74357477e311c7eafc149388a5f4f3ba"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.581563 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" event={"ID":"9f398c24-0435-49f5-915c-a54efeadea0c","Type":"ContainerStarted","Data":"d50e118282fe0f0c72f06721189015796ddd2f9781ceae03a507e05b89f39fcd"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.582515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" event={"ID":"ac49881b-2d45-4ba0-8b60-8461a2d5f35b","Type":"ContainerStarted","Data":"7d02ca8c8cc5e7956389bfb3b2736b958ff72e5511096562107e2a4bd32c1eca"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.583533 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" event={"ID":"79a719e7-5bce-424a-b837-ab0bca45936e","Type":"ContainerStarted","Data":"dc2fc9527b05ff054d153cf63751a9e30e681038cafba634b9f8deba506ede15"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.586315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" event={"ID":"d59f6f6b-419b-4ce3-adb6-11381aed6a5d","Type":"ContainerStarted","Data":"352436457f49531faf0232feabcea8aa84a2cb21c1531fb4fe4def804644d42a"} Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.813415 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.820256 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.828959 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.840994 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.842352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.842421 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw"] Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.846534 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst"] Oct 04 03:57:01 crc kubenswrapper[4726]: W1004 03:57:01.854183 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cb54ad2_85db_4291_87dc_4f0b9ea688c0.slice/crio-fdfea7693a0bedc8bba98728a223d6e1539183dd58a53b5823d905d4aeb18d9f WatchSource:0}: Error finding container fdfea7693a0bedc8bba98728a223d6e1539183dd58a53b5823d905d4aeb18d9f: Status 404 returned error can't find the container with id fdfea7693a0bedc8bba98728a223d6e1539183dd58a53b5823d905d4aeb18d9f Oct 04 03:57:01 crc kubenswrapper[4726]: W1004 03:57:01.856051 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e47df33_2621_4cfe_9c83_4d4bd1d616aa.slice/crio-58ddb60c60fa2693693c52d66537230608c81a6d6c9013ee4e5a619442ef1a2d WatchSource:0}: Error finding container 58ddb60c60fa2693693c52d66537230608c81a6d6c9013ee4e5a619442ef1a2d: Status 404 returned error can't find the container with id 58ddb60c60fa2693693c52d66537230608c81a6d6c9013ee4e5a619442ef1a2d Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.861345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb"] Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.869528 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tk224,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-5f7c849b98-mfsl5_openstack-operators(1e47df33-2621-4cfe-9c83-4d4bd1d616aa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.869846 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vlsdb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-4v7k8_openstack-operators(392e7503-f32e-4e8f-ac07-d75844253ea1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:01 crc kubenswrapper[4726]: I1004 03:57:01.869988 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm"] Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.870555 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6wssp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-lsxtw_openstack-operators(efa34c5f-e531-4319-b5fb-7497c4d026f0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:01 crc kubenswrapper[4726]: W1004 03:57:01.872425 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05b7da99_96e2_40be_8fc1_f847b8e7a643.slice/crio-4bb8d954d8ea993216eeb42734891859c7e4503c6dec21e3c7eddb366137cb99 WatchSource:0}: Error finding container 4bb8d954d8ea993216eeb42734891859c7e4503c6dec21e3c7eddb366137cb99: Status 404 returned error can't find the container with id 4bb8d954d8ea993216eeb42734891859c7e4503c6dec21e3c7eddb366137cb99 Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.876530 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-62jsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7c7fc454ff-mnndm_openstack-operators(05b7da99-96e2-40be-8fc1-f847b8e7a643): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:01 crc kubenswrapper[4726]: W1004 03:57:01.879743 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42949778_256b_4809_b2ce_d633428e649c.slice/crio-94eeef5cb422c20bc4b097fe062673d05acebb03523a2b350e2c6b9ede55037c WatchSource:0}: Error finding container 94eeef5cb422c20bc4b097fe062673d05acebb03523a2b350e2c6b9ede55037c: Status 404 returned error can't find the container with id 94eeef5cb422c20bc4b097fe062673d05acebb03523a2b350e2c6b9ede55037c Oct 04 03:57:01 crc kubenswrapper[4726]: E1004 03:57:01.882166 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dzd6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d4d74dd89-2j9rb_openstack-operators(42949778-256b-4809-b2ce-d633428e649c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.012338 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7"] Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.017515 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc"] Oct 04 03:57:02 crc kubenswrapper[4726]: W1004 03:57:02.028676 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod262abd5f_25d1_488f_b864_f9cc308a68f4.slice/crio-490d92986936337f0d7023b342c307237057738ba14ae6e66fbbc8c8ed806389 WatchSource:0}: Error finding container 490d92986936337f0d7023b342c307237057738ba14ae6e66fbbc8c8ed806389: Status 404 returned error can't find the container with id 490d92986936337f0d7023b342c307237057738ba14ae6e66fbbc8c8ed806389 Oct 04 03:57:02 crc kubenswrapper[4726]: W1004 03:57:02.031240 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod962ba8ee_fbcf_4ef5_b740_bd4a782d0881.slice/crio-3f7a439f485a8b2f00380bb6644b9ed20c687c9de19b8e4fc96e97d7030ceb53 WatchSource:0}: Error finding container 3f7a439f485a8b2f00380bb6644b9ed20c687c9de19b8e4fc96e97d7030ceb53: Status 404 returned error can't find the container with id 3f7a439f485a8b2f00380bb6644b9ed20c687c9de19b8e4fc96e97d7030ceb53 Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.031765 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b5xzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-kn2pc_openstack-operators(262abd5f-25d1-488f-b864-f9cc308a68f4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.033287 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c8lv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7_openstack-operators(962ba8ee-fbcf-4ef5-b740-bd4a782d0881): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.034371 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" podUID="962ba8ee-fbcf-4ef5-b740-bd4a782d0881" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.122108 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt"] Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.122531 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" podUID="1e47df33-2621-4cfe-9c83-4d4bd1d616aa" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.126380 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" podUID="42949778-256b-4809-b2ce-d633428e649c" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.141491 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" podUID="392e7503-f32e-4e8f-ac07-d75844253ea1" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.143173 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" podUID="05b7da99-96e2-40be-8fc1-f847b8e7a643" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.172635 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" podUID="efa34c5f-e531-4319-b5fb-7497c4d026f0" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.235051 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" podUID="262abd5f-25d1-488f-b864-f9cc308a68f4" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.452479 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.464007 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aaa6fd43-d6ed-4d86-9827-27d2529499e3-cert\") pod \"openstack-operator-controller-manager-ff7c6f545-hgqmd\" (UID: \"aaa6fd43-d6ed-4d86-9827-27d2529499e3\") " pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.594284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" event={"ID":"6508ae63-4f75-496f-b7b6-5dcd8a0cb008","Type":"ContainerStarted","Data":"ad53b6ae753c51a6825c08947b1cd23efda51f429924ec4ef60712b884a32950"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.596598 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" event={"ID":"ea261e29-40c5-4a26-a02f-8a09e0c7320d","Type":"ContainerStarted","Data":"f62c10a691b589a87b9c576903b8bb7f54292fb9f588ac90db39ee8b0aad557f"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.600005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" event={"ID":"1e47df33-2621-4cfe-9c83-4d4bd1d616aa","Type":"ContainerStarted","Data":"2253b58443a9002ca5d3f2323fc1f74f58e11be656bda0e7808d8ee8ef42733b"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.600028 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" event={"ID":"1e47df33-2621-4cfe-9c83-4d4bd1d616aa","Type":"ContainerStarted","Data":"58ddb60c60fa2693693c52d66537230608c81a6d6c9013ee4e5a619442ef1a2d"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.603457 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" event={"ID":"262abd5f-25d1-488f-b864-f9cc308a68f4","Type":"ContainerStarted","Data":"f404e8e9092df74d44233c5f5893cded8165e2f98c100d7ca4035f4564eb228b"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.603517 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" event={"ID":"262abd5f-25d1-488f-b864-f9cc308a68f4","Type":"ContainerStarted","Data":"490d92986936337f0d7023b342c307237057738ba14ae6e66fbbc8c8ed806389"} Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.605346 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" podUID="1e47df33-2621-4cfe-9c83-4d4bd1d616aa" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.605618 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" podUID="262abd5f-25d1-488f-b864-f9cc308a68f4" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.608004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" event={"ID":"392e7503-f32e-4e8f-ac07-d75844253ea1","Type":"ContainerStarted","Data":"5045a45402723bf868cbff33abf6eefbe16d576ccd83574719628aa112eb4216"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.608041 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" event={"ID":"392e7503-f32e-4e8f-ac07-d75844253ea1","Type":"ContainerStarted","Data":"cd0dbd9b02c9cd46fa9a90a1c8d22063e0acc6361be9c976166b72c9fc713424"} Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.609447 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" podUID="392e7503-f32e-4e8f-ac07-d75844253ea1" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.610448 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" event={"ID":"6b245f13-4e4b-4a41-b148-482d581e2c5c","Type":"ContainerStarted","Data":"eba6336ea6903517a3579c5afbf754821cd943e6a5045a107f56ddcde6940bd1"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.612399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" event={"ID":"05b7da99-96e2-40be-8fc1-f847b8e7a643","Type":"ContainerStarted","Data":"38a23a25b9d7233336618948298a8272b9cc32414eac12f8611262397c70f8f0"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.612434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" event={"ID":"05b7da99-96e2-40be-8fc1-f847b8e7a643","Type":"ContainerStarted","Data":"4bb8d954d8ea993216eeb42734891859c7e4503c6dec21e3c7eddb366137cb99"} Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.632578 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" podUID="05b7da99-96e2-40be-8fc1-f847b8e7a643" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.649905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.659455 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" event={"ID":"efa34c5f-e531-4319-b5fb-7497c4d026f0","Type":"ContainerStarted","Data":"4ee49b5d511629a9eabfa373c64b638577df48d5221e03cb61dbb96e7e2fa1e2"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.659503 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" event={"ID":"efa34c5f-e531-4319-b5fb-7497c4d026f0","Type":"ContainerStarted","Data":"a1d786fe580e8d8c8b74f3375e5af3931f3d8744a39e2ad433487f390c2a03f0"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.662606 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" event={"ID":"42949778-256b-4809-b2ce-d633428e649c","Type":"ContainerStarted","Data":"83195631d723d2fe63b81aef4fd6c20879ad0b2017d3a7638c6a9889fc34d726"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.662637 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" event={"ID":"42949778-256b-4809-b2ce-d633428e649c","Type":"ContainerStarted","Data":"94eeef5cb422c20bc4b097fe062673d05acebb03523a2b350e2c6b9ede55037c"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.664025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" event={"ID":"d55e66e5-0c46-4df2-909c-fc739a26c5cc","Type":"ContainerStarted","Data":"ac2ccd848e5d1c87e82ca3db8762784bbabd11b4b3c00a83ac613026b4d89f8f"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.665535 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" event={"ID":"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828","Type":"ContainerStarted","Data":"e3e776371a416af5d32f901095d2e2783eacac00e0a8a33f83039ad2344b2090"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.666561 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" event={"ID":"cadfc4b1-9327-4a55-8544-44422572319c","Type":"ContainerStarted","Data":"234ae37883e375050b6034614e95f15137b540e0739739acae7b1c123618e99f"} Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.667472 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" event={"ID":"962ba8ee-fbcf-4ef5-b740-bd4a782d0881","Type":"ContainerStarted","Data":"3f7a439f485a8b2f00380bb6644b9ed20c687c9de19b8e4fc96e97d7030ceb53"} Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.702357 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" podUID="42949778-256b-4809-b2ce-d633428e649c" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.702538 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" podUID="962ba8ee-fbcf-4ef5-b740-bd4a782d0881" Oct 04 03:57:02 crc kubenswrapper[4726]: E1004 03:57:02.702573 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" podUID="efa34c5f-e531-4319-b5fb-7497c4d026f0" Oct 04 03:57:02 crc kubenswrapper[4726]: I1004 03:57:02.725949 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" event={"ID":"5cb54ad2-85db-4291-87dc-4f0b9ea688c0","Type":"ContainerStarted","Data":"fdfea7693a0bedc8bba98728a223d6e1539183dd58a53b5823d905d4aeb18d9f"} Oct 04 03:57:03 crc kubenswrapper[4726]: I1004 03:57:03.274068 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd"] Oct 04 03:57:03 crc kubenswrapper[4726]: W1004 03:57:03.303655 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaa6fd43_d6ed_4d86_9827_27d2529499e3.slice/crio-d9456b7e5dc2fc9da88f75630df1b87f2578ec24695baf37a042a42341806e24 WatchSource:0}: Error finding container d9456b7e5dc2fc9da88f75630df1b87f2578ec24695baf37a042a42341806e24: Status 404 returned error can't find the container with id d9456b7e5dc2fc9da88f75630df1b87f2578ec24695baf37a042a42341806e24 Oct 04 03:57:03 crc kubenswrapper[4726]: I1004 03:57:03.748320 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" event={"ID":"aaa6fd43-d6ed-4d86-9827-27d2529499e3","Type":"ContainerStarted","Data":"32f8afb820f6ab10538e23390f70ba93c5e0a4bc3458c4aab1ce59be54da549f"} Oct 04 03:57:03 crc kubenswrapper[4726]: I1004 03:57:03.748618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" event={"ID":"aaa6fd43-d6ed-4d86-9827-27d2529499e3","Type":"ContainerStarted","Data":"d9456b7e5dc2fc9da88f75630df1b87f2578ec24695baf37a042a42341806e24"} Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756350 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" podUID="962ba8ee-fbcf-4ef5-b740-bd4a782d0881" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756357 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" podUID="42949778-256b-4809-b2ce-d633428e649c" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756393 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" podUID="262abd5f-25d1-488f-b864-f9cc308a68f4" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756414 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" podUID="392e7503-f32e-4e8f-ac07-d75844253ea1" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756412 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" podUID="efa34c5f-e531-4319-b5fb-7497c4d026f0" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.756402 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" podUID="05b7da99-96e2-40be-8fc1-f847b8e7a643" Oct 04 03:57:03 crc kubenswrapper[4726]: E1004 03:57:03.757181 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:f9fc3cf4084a325d7f5f9773bfcc2b839ccff1c72e61fdd8f410a7ef46497f75\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" podUID="1e47df33-2621-4cfe-9c83-4d4bd1d616aa" Oct 04 03:57:14 crc kubenswrapper[4726]: E1004 03:57:14.063021 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73" Oct 04 03:57:14 crc kubenswrapper[4726]: E1004 03:57:14.063585 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-586bv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-54876c876f-fvw5p_openstack-operators(5e5f2785-db94-4fdc-9cb2-10b1ce940694): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:57:14 crc kubenswrapper[4726]: E1004 03:57:14.600800 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" podUID="5e5f2785-db94-4fdc-9cb2-10b1ce940694" Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.844593 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" event={"ID":"5cb54ad2-85db-4291-87dc-4f0b9ea688c0","Type":"ContainerStarted","Data":"3f4551014f28499c0448f5e28469fd40f9b803dfdcf3fa115d2b23f8a26de427"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.845866 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" event={"ID":"6508ae63-4f75-496f-b7b6-5dcd8a0cb008","Type":"ContainerStarted","Data":"4c203980790a070ec6605820aedd5ad010b9ed6e9faf1fcfd381fd2eadcee60d"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.863134 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" event={"ID":"cadfc4b1-9327-4a55-8544-44422572319c","Type":"ContainerStarted","Data":"51be0eb5da3491cf6136e0c90b89a948ae3027697adf891ad50792ed3d1eb75d"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.874400 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" event={"ID":"aaa6fd43-d6ed-4d86-9827-27d2529499e3","Type":"ContainerStarted","Data":"18dc6d4eaf17b04bc9bbc1b60c65c3b847e9e8790df32f2251ce1038d42844b4"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.874620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.882296 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.883884 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" event={"ID":"79a719e7-5bce-424a-b837-ab0bca45936e","Type":"ContainerStarted","Data":"d1122fd70c46223804e60351c9f11d5e6e6109dfc82b6aa73f7e16e80edaf013"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.904013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" event={"ID":"d59f6f6b-419b-4ce3-adb6-11381aed6a5d","Type":"ContainerStarted","Data":"28efa84ee088cb5452d586f4260757d97e161c0950d85a305a147dd51d30e60a"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.909054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" event={"ID":"5e5f2785-db94-4fdc-9cb2-10b1ce940694","Type":"ContainerStarted","Data":"56667b19c5978281ef5676b1fb7cfb37b51888eb0251f467bf1118e210af5d3b"} Oct 04 03:57:14 crc kubenswrapper[4726]: E1004 03:57:14.913122 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" podUID="5e5f2785-db94-4fdc-9cb2-10b1ce940694" Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.924371 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" event={"ID":"9f398c24-0435-49f5-915c-a54efeadea0c","Type":"ContainerStarted","Data":"c6059a8ada620031136ebbc4b898125c6091146b906033e5066ce8f17356f9ad"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.924776 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-ff7c6f545-hgqmd" podStartSLOduration=14.924754484 podStartE2EDuration="14.924754484s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:57:14.923317956 +0000 UTC m=+1009.097941169" watchObservedRunningTime="2025-10-04 03:57:14.924754484 +0000 UTC m=+1009.099377697" Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.931417 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" event={"ID":"ea261e29-40c5-4a26-a02f-8a09e0c7320d","Type":"ContainerStarted","Data":"fae78875cc0e25d34f81722a52843c5ca08b8ab603a6d6110eea294e9c1a6684"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.946211 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" event={"ID":"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3","Type":"ContainerStarted","Data":"832068422d513677469cd1c4713411b48a6bbea7bdc2d407c0cf44cfc8adcea9"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.962712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" event={"ID":"d55e66e5-0c46-4df2-909c-fc739a26c5cc","Type":"ContainerStarted","Data":"56b285257768a3293a3b7e7e9e8b082f8dc28d94c63b4842983eee24bcccf081"} Oct 04 03:57:14 crc kubenswrapper[4726]: I1004 03:57:14.976740 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" event={"ID":"a520bf75-9223-4209-87d4-248ce3e0d190","Type":"ContainerStarted","Data":"d64d715e39566d49c7dea813ddd41ce15062beafb92d5b2bc5a33bf759a35517"} Oct 04 03:57:15 crc kubenswrapper[4726]: I1004 03:57:15.017843 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" event={"ID":"ac49881b-2d45-4ba0-8b60-8461a2d5f35b","Type":"ContainerStarted","Data":"dff15cf19ff5ffdd461cca172cec8543b706d28869a4e83f10ace5b9870e8ef0"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.025823 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" event={"ID":"d55e66e5-0c46-4df2-909c-fc739a26c5cc","Type":"ContainerStarted","Data":"50197c804014ab7f8ab227bb1ee4634d7e1cf2869f9f5f2f1842c65aa12c94bb"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.026128 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.027019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" event={"ID":"a520bf75-9223-4209-87d4-248ce3e0d190","Type":"ContainerStarted","Data":"ff00b5ac57109d367af46154de199b7c4c901b74f2da5793cbcf0ee64eccfcc4"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.027150 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.031089 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" event={"ID":"79a719e7-5bce-424a-b837-ab0bca45936e","Type":"ContainerStarted","Data":"0df80acc1fd992375fa21835fe7f440d053960e8e6ae6ea11ed1bd7ae48fd419"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.031264 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.032299 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" event={"ID":"9f398c24-0435-49f5-915c-a54efeadea0c","Type":"ContainerStarted","Data":"2cc8e29c81a9a242060ead0628e1d4a77e316a0ee9e3a3579b64d76648fb5a39"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.032656 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.034208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" event={"ID":"6508ae63-4f75-496f-b7b6-5dcd8a0cb008","Type":"ContainerStarted","Data":"654da21c625b3edcd93a9ae27a27ef5e0b9aee94bd99eb02d0c0f269552510c9"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.034534 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.038510 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" event={"ID":"f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3","Type":"ContainerStarted","Data":"e9f7d2a27c5f3ae7fe55cf1aa4502cbe96ec0ebfaede924eed4167c18a21a564"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.038936 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.042366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" event={"ID":"5cb54ad2-85db-4291-87dc-4f0b9ea688c0","Type":"ContainerStarted","Data":"c5095bc643cb677eceb093ef731634b8b717979bab734e2ccf3593d6b710d626"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.042688 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.045014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" event={"ID":"cadfc4b1-9327-4a55-8544-44422572319c","Type":"ContainerStarted","Data":"744ef51afca572831c4ed9e0397a5abf8fceab62e1f7a347e0da897c5b07e25a"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.045370 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.048400 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" event={"ID":"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828","Type":"ContainerStarted","Data":"97a793bcedc10941e494516b3aedd21fd080cd95e0cc6e6a1b0bfb6c886d0133"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.048442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" event={"ID":"e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828","Type":"ContainerStarted","Data":"9d5a2fcba0fd4d1fbdcb5118ba0f0adfeac790a7c622acd58041ae2dc87f6ffc"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.049082 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.049652 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" podStartSLOduration=3.501166859 podStartE2EDuration="16.049639676s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.581463529 +0000 UTC m=+995.756086742" lastFinishedPulling="2025-10-04 03:57:14.129936356 +0000 UTC m=+1008.304559559" observedRunningTime="2025-10-04 03:57:16.047673514 +0000 UTC m=+1010.222296727" watchObservedRunningTime="2025-10-04 03:57:16.049639676 +0000 UTC m=+1010.224262879" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.061467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" event={"ID":"ac49881b-2d45-4ba0-8b60-8461a2d5f35b","Type":"ContainerStarted","Data":"4da3582dc73fcd3728bb4cbabb5d107ddacbea5035dd31480948af4881a818a2"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.061613 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.068895 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" podStartSLOduration=3.05234634 podStartE2EDuration="16.068883472s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.071684798 +0000 UTC m=+995.246308001" lastFinishedPulling="2025-10-04 03:57:14.08822191 +0000 UTC m=+1008.262845133" observedRunningTime="2025-10-04 03:57:16.063091829 +0000 UTC m=+1010.237715042" watchObservedRunningTime="2025-10-04 03:57:16.068883472 +0000 UTC m=+1010.243506685" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.077485 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" podStartSLOduration=4.484126433 podStartE2EDuration="17.077469287s" podCreationTimestamp="2025-10-04 03:56:59 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.553676651 +0000 UTC m=+995.728299864" lastFinishedPulling="2025-10-04 03:57:14.147019505 +0000 UTC m=+1008.321642718" observedRunningTime="2025-10-04 03:57:16.074658133 +0000 UTC m=+1010.249281346" watchObservedRunningTime="2025-10-04 03:57:16.077469287 +0000 UTC m=+1010.252092500" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.081743 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" event={"ID":"6b245f13-4e4b-4a41-b148-482d581e2c5c","Type":"ContainerStarted","Data":"d8cefa30540190d1710df37ca512f1d4815f52b4d40697190b489875cfc8fcd1"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.081787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" event={"ID":"6b245f13-4e4b-4a41-b148-482d581e2c5c","Type":"ContainerStarted","Data":"96a22f0de404102f5d76c4c9e967d0eebecbdd3d51eefb6a7f484e9649a3cdca"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.081836 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.084239 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" event={"ID":"d59f6f6b-419b-4ce3-adb6-11381aed6a5d","Type":"ContainerStarted","Data":"4fb27a7012e57cae04330eff14977881a821ac5e257eb6d9a7e42d19a529738a"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.084815 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.086471 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" event={"ID":"ea261e29-40c5-4a26-a02f-8a09e0c7320d","Type":"ContainerStarted","Data":"a019ef302affe4836e2a1d96b92a9f3b03dafbbf6b1817003040e0bd7ea69ad5"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.086783 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.088221 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" event={"ID":"488fc85d-f45f-4d8d-93e2-b58ba9306a76","Type":"ContainerStarted","Data":"af0a914d71b427396b198fbb1ef429e634d18a33ee9c0f8d048a775f0e8f0790"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.088275 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" event={"ID":"488fc85d-f45f-4d8d-93e2-b58ba9306a76","Type":"ContainerStarted","Data":"42cbbfa4191f4753cce737ab694761d3d1e95f90aee219124cf7bdd4224d456d"} Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.088358 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:16 crc kubenswrapper[4726]: E1004 03:57:16.089345 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:4cba007c18be1ec9aac2ece7a5ce6444a94afd89f0fb032522811d5bdf5bee73\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" podUID="5e5f2785-db94-4fdc-9cb2-10b1ce940694" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.109576 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" podStartSLOduration=3.586294809 podStartE2EDuration="16.10955263s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.578350511 +0000 UTC m=+995.752973724" lastFinishedPulling="2025-10-04 03:57:14.101608332 +0000 UTC m=+1008.276231545" observedRunningTime="2025-10-04 03:57:16.093247682 +0000 UTC m=+1010.267870895" watchObservedRunningTime="2025-10-04 03:57:16.10955263 +0000 UTC m=+1010.284175843" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.129795 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" podStartSLOduration=4.103601808 podStartE2EDuration="17.129778432s" podCreationTimestamp="2025-10-04 03:56:59 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.080718627 +0000 UTC m=+995.255341840" lastFinishedPulling="2025-10-04 03:57:14.106895231 +0000 UTC m=+1008.281518464" observedRunningTime="2025-10-04 03:57:16.126562607 +0000 UTC m=+1010.301185820" watchObservedRunningTime="2025-10-04 03:57:16.129778432 +0000 UTC m=+1010.304401645" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.144834 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" podStartSLOduration=3.895670075 podStartE2EDuration="16.144815807s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.869394795 +0000 UTC m=+996.044018008" lastFinishedPulling="2025-10-04 03:57:14.118540527 +0000 UTC m=+1008.293163740" observedRunningTime="2025-10-04 03:57:16.142419334 +0000 UTC m=+1010.317042547" watchObservedRunningTime="2025-10-04 03:57:16.144815807 +0000 UTC m=+1010.319439020" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.160975 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" podStartSLOduration=3.930055511 podStartE2EDuration="16.160956581s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.86265156 +0000 UTC m=+996.037274773" lastFinishedPulling="2025-10-04 03:57:14.09355263 +0000 UTC m=+1008.268175843" observedRunningTime="2025-10-04 03:57:16.157565652 +0000 UTC m=+1010.332188865" watchObservedRunningTime="2025-10-04 03:57:16.160956581 +0000 UTC m=+1010.335579794" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.188999 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" podStartSLOduration=3.595582563 podStartE2EDuration="16.188972688s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.553672741 +0000 UTC m=+995.728295954" lastFinishedPulling="2025-10-04 03:57:14.147062846 +0000 UTC m=+1008.321686079" observedRunningTime="2025-10-04 03:57:16.177705971 +0000 UTC m=+1010.352329184" watchObservedRunningTime="2025-10-04 03:57:16.188972688 +0000 UTC m=+1010.363595901" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.209185 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" podStartSLOduration=4.232426938 podStartE2EDuration="16.209166508s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:02.147836675 +0000 UTC m=+996.322459888" lastFinishedPulling="2025-10-04 03:57:14.124576225 +0000 UTC m=+1008.299199458" observedRunningTime="2025-10-04 03:57:16.207655129 +0000 UTC m=+1010.382278342" watchObservedRunningTime="2025-10-04 03:57:16.209166508 +0000 UTC m=+1010.383789721" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.247337 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" podStartSLOduration=3.594011746 podStartE2EDuration="16.247311111s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.516473438 +0000 UTC m=+995.691096651" lastFinishedPulling="2025-10-04 03:57:14.169772793 +0000 UTC m=+1008.344396016" observedRunningTime="2025-10-04 03:57:16.233866597 +0000 UTC m=+1010.408489820" watchObservedRunningTime="2025-10-04 03:57:16.247311111 +0000 UTC m=+1010.421934324" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.282477 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" podStartSLOduration=4.009387709 podStartE2EDuration="16.282458524s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.815075883 +0000 UTC m=+995.989699096" lastFinishedPulling="2025-10-04 03:57:14.088146698 +0000 UTC m=+1008.262769911" observedRunningTime="2025-10-04 03:57:16.281342885 +0000 UTC m=+1010.455966098" watchObservedRunningTime="2025-10-04 03:57:16.282458524 +0000 UTC m=+1010.457081737" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.304343 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" podStartSLOduration=3.776639999 podStartE2EDuration="16.304327799s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.553970547 +0000 UTC m=+995.728593760" lastFinishedPulling="2025-10-04 03:57:14.081658347 +0000 UTC m=+1008.256281560" observedRunningTime="2025-10-04 03:57:16.294550092 +0000 UTC m=+1010.469173305" watchObservedRunningTime="2025-10-04 03:57:16.304327799 +0000 UTC m=+1010.478951012" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.321901 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" podStartSLOduration=3.476800469 podStartE2EDuration="16.321887241s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.28851515 +0000 UTC m=+995.463138363" lastFinishedPulling="2025-10-04 03:57:14.133601912 +0000 UTC m=+1008.308225135" observedRunningTime="2025-10-04 03:57:16.318617285 +0000 UTC m=+1010.493240498" watchObservedRunningTime="2025-10-04 03:57:16.321887241 +0000 UTC m=+1010.496510454" Oct 04 03:57:16 crc kubenswrapper[4726]: I1004 03:57:16.344547 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" podStartSLOduration=4.076679044 podStartE2EDuration="16.344531136s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.873102894 +0000 UTC m=+996.047726107" lastFinishedPulling="2025-10-04 03:57:14.140954986 +0000 UTC m=+1008.315578199" observedRunningTime="2025-10-04 03:57:16.339852243 +0000 UTC m=+1010.514475456" watchObservedRunningTime="2025-10-04 03:57:16.344531136 +0000 UTC m=+1010.519154349" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.117315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" event={"ID":"05b7da99-96e2-40be-8fc1-f847b8e7a643","Type":"ContainerStarted","Data":"02c4b6e56bcea6a008b870f8b2d21476f47cf56387b66f35a53bfed157b3f323"} Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.118050 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.121906 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" event={"ID":"962ba8ee-fbcf-4ef5-b740-bd4a782d0881","Type":"ContainerStarted","Data":"2c9a7588a657e4712f103c3136a1c7dfce73d20e9f1ca5390ace759a906d9829"} Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.124026 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" event={"ID":"42949778-256b-4809-b2ce-d633428e649c","Type":"ContainerStarted","Data":"c9b78f26964ee773c2cdba0e024cf667b004522ef5e41be9e6112a28f40e6616"} Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.124238 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.126751 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" event={"ID":"262abd5f-25d1-488f-b864-f9cc308a68f4","Type":"ContainerStarted","Data":"146730e41a37af851b5836253348910b239d663f97445fd7f9c5119d2d1e22b0"} Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.126960 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.131271 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" event={"ID":"392e7503-f32e-4e8f-ac07-d75844253ea1","Type":"ContainerStarted","Data":"b504c30c22f4067db9e6ac2a5e5d866e684a5bacb283fd7baa22b1f55c1e0de8"} Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.131480 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.142829 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" podStartSLOduration=3.055684258 podStartE2EDuration="20.142807563s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.876379136 +0000 UTC m=+996.051002349" lastFinishedPulling="2025-10-04 03:57:18.963502451 +0000 UTC m=+1013.138125654" observedRunningTime="2025-10-04 03:57:20.139407604 +0000 UTC m=+1014.314030837" watchObservedRunningTime="2025-10-04 03:57:20.142807563 +0000 UTC m=+1014.317430776" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.159601 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" podStartSLOduration=3.091201658 podStartE2EDuration="20.159581544s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.882013021 +0000 UTC m=+996.056636234" lastFinishedPulling="2025-10-04 03:57:18.950392907 +0000 UTC m=+1013.125016120" observedRunningTime="2025-10-04 03:57:20.158701831 +0000 UTC m=+1014.333325054" watchObservedRunningTime="2025-10-04 03:57:20.159581544 +0000 UTC m=+1014.334204757" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.181953 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7" podStartSLOduration=3.288027658 podStartE2EDuration="20.181932001s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:02.032809621 +0000 UTC m=+996.207432954" lastFinishedPulling="2025-10-04 03:57:18.926714084 +0000 UTC m=+1013.101337297" observedRunningTime="2025-10-04 03:57:20.176718814 +0000 UTC m=+1014.351342037" watchObservedRunningTime="2025-10-04 03:57:20.181932001 +0000 UTC m=+1014.356555224" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.197765 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" podStartSLOduration=3.138071142 podStartE2EDuration="20.197741746s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.869750472 +0000 UTC m=+996.044373685" lastFinishedPulling="2025-10-04 03:57:18.929421076 +0000 UTC m=+1013.104044289" observedRunningTime="2025-10-04 03:57:20.192771085 +0000 UTC m=+1014.367394308" watchObservedRunningTime="2025-10-04 03:57:20.197741746 +0000 UTC m=+1014.372364969" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.217043 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" podStartSLOduration=3.328230103 podStartE2EDuration="20.217025453s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:02.031634339 +0000 UTC m=+996.206257552" lastFinishedPulling="2025-10-04 03:57:18.920429689 +0000 UTC m=+1013.095052902" observedRunningTime="2025-10-04 03:57:20.212974186 +0000 UTC m=+1014.387597409" watchObservedRunningTime="2025-10-04 03:57:20.217025453 +0000 UTC m=+1014.391648666" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.364990 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-tzr27" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.448885 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-cjx6d" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.472936 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-lqs6l" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.564834 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-v72j6" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.580828 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-wwdj2" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.596603 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-6dpjp" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.649786 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-s67gj" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.674509 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-lqsxr" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.718199 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-gtshn" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.809767 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-zkgst" Oct 04 03:57:20 crc kubenswrapper[4726]: I1004 03:57:20.867545 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-8pnpc" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.032911 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-hv9lz" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.112321 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7ddb876846-kgv8l" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.139776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" event={"ID":"efa34c5f-e531-4319-b5fb-7497c4d026f0","Type":"ContainerStarted","Data":"1a405fea1d7923ebf6cb07eb7ca39e57a236543ca4482b25d35e105f806dd4e6"} Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.140028 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.142502 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" event={"ID":"1e47df33-2621-4cfe-9c83-4d4bd1d616aa","Type":"ContainerStarted","Data":"765aa81c4aa7b4582de630fff6aeca1b96852a0ff076ab9898be4df27c799c63"} Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.173668 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" podStartSLOduration=2.296958953 podStartE2EDuration="21.173645233s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.870473015 +0000 UTC m=+996.045096228" lastFinishedPulling="2025-10-04 03:57:20.747159295 +0000 UTC m=+1014.921782508" observedRunningTime="2025-10-04 03:57:21.168316803 +0000 UTC m=+1015.342940016" watchObservedRunningTime="2025-10-04 03:57:21.173645233 +0000 UTC m=+1015.348268446" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.458125 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt" Oct 04 03:57:21 crc kubenswrapper[4726]: I1004 03:57:21.508715 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" podStartSLOduration=3.63325655 podStartE2EDuration="22.508688228s" podCreationTimestamp="2025-10-04 03:56:59 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.869418266 +0000 UTC m=+996.044041479" lastFinishedPulling="2025-10-04 03:57:20.744849944 +0000 UTC m=+1014.919473157" observedRunningTime="2025-10-04 03:57:21.191484032 +0000 UTC m=+1015.366107245" watchObservedRunningTime="2025-10-04 03:57:21.508688228 +0000 UTC m=+1015.683311481" Oct 04 03:57:28 crc kubenswrapper[4726]: I1004 03:57:28.198037 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" event={"ID":"5e5f2785-db94-4fdc-9cb2-10b1ce940694","Type":"ContainerStarted","Data":"d4976ad819398e7e0e589d24897c692a86806904a8b28a7d44a8d3ec4459a816"} Oct 04 03:57:28 crc kubenswrapper[4726]: I1004 03:57:28.198656 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:57:28 crc kubenswrapper[4726]: I1004 03:57:28.229749 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" podStartSLOduration=1.737101513 podStartE2EDuration="28.229726546s" podCreationTimestamp="2025-10-04 03:57:00 +0000 UTC" firstStartedPulling="2025-10-04 03:57:01.082584252 +0000 UTC m=+995.257207465" lastFinishedPulling="2025-10-04 03:57:27.575209295 +0000 UTC m=+1021.749832498" observedRunningTime="2025-10-04 03:57:28.217140475 +0000 UTC m=+1022.391763718" watchObservedRunningTime="2025-10-04 03:57:28.229726546 +0000 UTC m=+1022.404349789" Oct 04 03:57:30 crc kubenswrapper[4726]: I1004 03:57:30.619775 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:30 crc kubenswrapper[4726]: I1004 03:57:30.621679 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-mfsl5" Oct 04 03:57:30 crc kubenswrapper[4726]: I1004 03:57:30.780521 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mnndm" Oct 04 03:57:30 crc kubenswrapper[4726]: I1004 03:57:30.983503 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-lsxtw" Oct 04 03:57:31 crc kubenswrapper[4726]: I1004 03:57:31.023916 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-2j9rb" Oct 04 03:57:31 crc kubenswrapper[4726]: I1004 03:57:31.096764 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-4v7k8" Oct 04 03:57:31 crc kubenswrapper[4726]: I1004 03:57:31.115436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-kn2pc" Oct 04 03:57:40 crc kubenswrapper[4726]: I1004 03:57:40.497416 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-fvw5p" Oct 04 03:58:01 crc kubenswrapper[4726]: I1004 03:58:01.985865 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.002002 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.002137 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.005419 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.005654 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-np2h2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.009340 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.009356 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.022219 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.022332 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stn5j\" (UniqueName: \"kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.092696 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.097688 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.099597 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.101417 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.125354 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.125502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sggsw\" (UniqueName: \"kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.125679 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stn5j\" (UniqueName: \"kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.125716 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.125737 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.126992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.162174 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stn5j\" (UniqueName: \"kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j\") pod \"dnsmasq-dns-675f4bcbfc-lxh56\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.226888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sggsw\" (UniqueName: \"kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.226946 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.226984 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.227844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.228339 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.246355 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sggsw\" (UniqueName: \"kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw\") pod \"dnsmasq-dns-78dd6ddcc-vtqc2\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.327343 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.433220 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.657605 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:02 crc kubenswrapper[4726]: W1004 03:58:02.660702 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod730d38e1_c902_4ff0_af36_8434cc2b94fd.slice/crio-3e76af8c25fced04341960ffacf3801fd94fbd723065edc5e003f5bbe3546cea WatchSource:0}: Error finding container 3e76af8c25fced04341960ffacf3801fd94fbd723065edc5e003f5bbe3546cea: Status 404 returned error can't find the container with id 3e76af8c25fced04341960ffacf3801fd94fbd723065edc5e003f5bbe3546cea Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.663184 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 03:58:02 crc kubenswrapper[4726]: I1004 03:58:02.749332 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:02 crc kubenswrapper[4726]: W1004 03:58:02.753521 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa16927f_b443_4ed4_9988_aedb52dcd8a9.slice/crio-b528a37c562fac0ef72c1e78f56a8807dee3aafd661381ad159ea7ae74561a7f WatchSource:0}: Error finding container b528a37c562fac0ef72c1e78f56a8807dee3aafd661381ad159ea7ae74561a7f: Status 404 returned error can't find the container with id b528a37c562fac0ef72c1e78f56a8807dee3aafd661381ad159ea7ae74561a7f Oct 04 03:58:03 crc kubenswrapper[4726]: I1004 03:58:03.536964 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" event={"ID":"730d38e1-c902-4ff0-af36-8434cc2b94fd","Type":"ContainerStarted","Data":"3e76af8c25fced04341960ffacf3801fd94fbd723065edc5e003f5bbe3546cea"} Oct 04 03:58:03 crc kubenswrapper[4726]: I1004 03:58:03.538845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" event={"ID":"fa16927f-b443-4ed4-9988-aedb52dcd8a9","Type":"ContainerStarted","Data":"b528a37c562fac0ef72c1e78f56a8807dee3aafd661381ad159ea7ae74561a7f"} Oct 04 03:58:04 crc kubenswrapper[4726]: I1004 03:58:04.950807 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:04 crc kubenswrapper[4726]: I1004 03:58:04.976234 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:04 crc kubenswrapper[4726]: I1004 03:58:04.977671 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.000333 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.179976 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfg8p\" (UniqueName: \"kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.180086 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.180173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.273346 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.281329 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.281387 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.281486 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfg8p\" (UniqueName: \"kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.282653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.282660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.304778 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.306586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.310494 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfg8p\" (UniqueName: \"kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p\") pod \"dnsmasq-dns-666b6646f7-wjrvs\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.313749 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.486139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk7hk\" (UniqueName: \"kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.486183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.486234 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.587471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk7hk\" (UniqueName: \"kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.587520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.587569 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.588358 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.589526 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.596585 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.610550 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk7hk\" (UniqueName: \"kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk\") pod \"dnsmasq-dns-57d769cc4f-ls5ds\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.648692 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:05 crc kubenswrapper[4726]: I1004 03:58:05.914801 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.152267 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.159143 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.159232 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.161657 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.161870 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.161992 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.162228 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.162463 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.163223 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.163269 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p9vjw" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.165245 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:58:06 crc kubenswrapper[4726]: W1004 03:58:06.167196 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4c1331_eccd_49e9_be27_242954c93434.slice/crio-418d0f8fec15598b0bb61232d7869f484a3dad0c0a13919121b5a932e6101f13 WatchSource:0}: Error finding container 418d0f8fec15598b0bb61232d7869f484a3dad0c0a13919121b5a932e6101f13: Status 404 returned error can't find the container with id 418d0f8fec15598b0bb61232d7869f484a3dad0c0a13919121b5a932e6101f13 Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298281 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298402 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298425 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298519 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298549 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298636 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwldb\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298698 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298766 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298838 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.298901 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400075 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwldb\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400135 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400177 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400225 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400245 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400262 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400287 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400339 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.400844 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.401497 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.401553 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.402331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.402546 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.403468 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.404586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.405379 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.413865 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.416828 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.418769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwldb\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.422804 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.434765 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.435452 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.440529 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.440767 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.440518 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.441189 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.441550 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.441621 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2zxzd" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.442582 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.449963 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.490893 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.579238 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" event={"ID":"254c8ce5-3f94-47f4-8f8a-e792235c5378","Type":"ContainerStarted","Data":"5a6cda46ef4a079ffaeef5f546c74bca089c9d02536e6e380ea6fc0796629914"} Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.586738 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" event={"ID":"9a4c1331-eccd-49e9-be27-242954c93434","Type":"ContainerStarted","Data":"418d0f8fec15598b0bb61232d7869f484a3dad0c0a13919121b5a932e6101f13"} Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603411 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603474 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603524 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm2kg\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603615 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603669 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603697 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.603766 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705273 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705323 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705367 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705394 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm2kg\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705486 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705504 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705530 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705552 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705549 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.705879 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.706584 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.706762 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.707588 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.707669 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.711764 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.711954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.712180 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.712523 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.722777 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm2kg\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.723769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.797192 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:58:06 crc kubenswrapper[4726]: I1004 03:58:06.920816 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 03:58:07 crc kubenswrapper[4726]: I1004 03:58:07.204734 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.047268 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.049401 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.053352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.053955 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.054868 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-lg2f2" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.055162 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.056029 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.058456 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.060429 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.061932 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.065136 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.067036 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.068178 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.068622 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6pf4j" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.069313 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.069366 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149698 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149744 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-kolla-config\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149851 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149899 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-config-data-default\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.149993 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95jbz\" (UniqueName: \"kubernetes.io/projected/d148624b-4c10-4a20-b428-902273d70ac0-kube-api-access-95jbz\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150019 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55tp9\" (UniqueName: \"kubernetes.io/projected/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kube-api-access-55tp9\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150057 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d148624b-4c10-4a20-b428-902273d70ac0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150082 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150139 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150179 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150201 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-secrets\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150234 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150384 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.150430 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251722 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-secrets\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251752 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251821 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251840 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-kolla-config\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251903 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251919 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251937 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-config-data-default\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251952 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251974 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jbz\" (UniqueName: \"kubernetes.io/projected/d148624b-4c10-4a20-b428-902273d70ac0-kube-api-access-95jbz\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.251995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55tp9\" (UniqueName: \"kubernetes.io/projected/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kube-api-access-55tp9\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.252020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d148624b-4c10-4a20-b428-902273d70ac0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.252037 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.252057 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.252930 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.253275 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.254225 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.256861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-kolla-config\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.256940 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.257087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d148624b-4c10-4a20-b428-902273d70ac0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.257976 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.258499 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.258753 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-config-data-default\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.258803 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d148624b-4c10-4a20-b428-902273d70ac0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.259157 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.259639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.260051 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.268691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.268952 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d148624b-4c10-4a20-b428-902273d70ac0-secrets\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.269586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jbz\" (UniqueName: \"kubernetes.io/projected/d148624b-4c10-4a20-b428-902273d70ac0-kube-api-access-95jbz\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.271303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.272637 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55tp9\" (UniqueName: \"kubernetes.io/projected/ed4e96c4-5868-4fd7-970b-ca7c74b56b1a-kube-api-access-55tp9\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.299647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"d148624b-4c10-4a20-b428-902273d70ac0\") " pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.302170 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a\") " pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.374652 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.389913 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.683390 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.684552 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.689943 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.690023 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-87dbk" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.690127 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.697345 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.763347 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-config-data\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.763418 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.763494 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rqwb\" (UniqueName: \"kubernetes.io/projected/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kube-api-access-8rqwb\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.763512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kolla-config\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.763658 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.865382 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.865470 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-config-data\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.865514 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.865567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rqwb\" (UniqueName: \"kubernetes.io/projected/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kube-api-access-8rqwb\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.865591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kolla-config\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.866604 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kolla-config\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.866699 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-config-data\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.870865 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.886138 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: I1004 03:58:09.893898 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rqwb\" (UniqueName: \"kubernetes.io/projected/5b1149f2-4d8f-4b1e-b5ec-b13f9833c706-kube-api-access-8rqwb\") pod \"memcached-0\" (UID: \"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706\") " pod="openstack/memcached-0" Oct 04 03:58:09 crc kubenswrapper[4726]: W1004 03:58:09.992072 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7f9fe7b_1700_4842_bda1_354de76c325c.slice/crio-989a8f262ff96d221b13e88780dbd27de380ade6d7dc2374a836190da0bb1306 WatchSource:0}: Error finding container 989a8f262ff96d221b13e88780dbd27de380ade6d7dc2374a836190da0bb1306: Status 404 returned error can't find the container with id 989a8f262ff96d221b13e88780dbd27de380ade6d7dc2374a836190da0bb1306 Oct 04 03:58:10 crc kubenswrapper[4726]: I1004 03:58:10.004953 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 03:58:10 crc kubenswrapper[4726]: W1004 03:58:10.604683 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf5efa33_a191_461b_baf3_9b12f2dbd4aa.slice/crio-f41135e4aa0ae60de7e78dcecd27e19bdd17457df8141c36c441d27e807386cf WatchSource:0}: Error finding container f41135e4aa0ae60de7e78dcecd27e19bdd17457df8141c36c441d27e807386cf: Status 404 returned error can't find the container with id f41135e4aa0ae60de7e78dcecd27e19bdd17457df8141c36c441d27e807386cf Oct 04 03:58:10 crc kubenswrapper[4726]: I1004 03:58:10.621319 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerStarted","Data":"989a8f262ff96d221b13e88780dbd27de380ade6d7dc2374a836190da0bb1306"} Oct 04 03:58:10 crc kubenswrapper[4726]: I1004 03:58:10.624082 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerStarted","Data":"f41135e4aa0ae60de7e78dcecd27e19bdd17457df8141c36c441d27e807386cf"} Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.725001 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.742026 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.742669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.747501 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-xcrn2" Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.804777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgb5g\" (UniqueName: \"kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g\") pod \"kube-state-metrics-0\" (UID: \"5d1c61fb-2030-40d7-a859-7f8c30983fc1\") " pod="openstack/kube-state-metrics-0" Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.906537 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgb5g\" (UniqueName: \"kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g\") pod \"kube-state-metrics-0\" (UID: \"5d1c61fb-2030-40d7-a859-7f8c30983fc1\") " pod="openstack/kube-state-metrics-0" Oct 04 03:58:11 crc kubenswrapper[4726]: I1004 03:58:11.923971 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgb5g\" (UniqueName: \"kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g\") pod \"kube-state-metrics-0\" (UID: \"5d1c61fb-2030-40d7-a859-7f8c30983fc1\") " pod="openstack/kube-state-metrics-0" Oct 04 03:58:12 crc kubenswrapper[4726]: I1004 03:58:12.065051 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.185330 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5kwdw"] Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.188037 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.190927 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-rbfdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.191427 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.191667 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.205516 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7724w"] Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.209541 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.212561 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw"] Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.240632 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7724w"] Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256489 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8675d591-017c-4479-bcb2-dbb70c2f9f93-scripts\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256520 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-log-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256543 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-ovn-controller-tls-certs\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256574 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256598 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-combined-ca-bundle\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.256636 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9qkg\" (UniqueName: \"kubernetes.io/projected/8675d591-017c-4479-bcb2-dbb70c2f9f93-kube-api-access-m9qkg\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358365 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-etc-ovs\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358442 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-run\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358477 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-log\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358496 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/178e673b-646c-44a4-b368-29d3fabce0c4-scripts\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8675d591-017c-4479-bcb2-dbb70c2f9f93-scripts\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-log-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358636 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-ovn-controller-tls-certs\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358742 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp8hk\" (UniqueName: \"kubernetes.io/projected/178e673b-646c-44a4-b368-29d3fabce0c4-kube-api-access-pp8hk\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358805 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-lib\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358911 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-combined-ca-bundle\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.358996 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9qkg\" (UniqueName: \"kubernetes.io/projected/8675d591-017c-4479-bcb2-dbb70c2f9f93-kube-api-access-m9qkg\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.359668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-log-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.359776 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run-ovn\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.359923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8675d591-017c-4479-bcb2-dbb70c2f9f93-var-run\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.361037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8675d591-017c-4479-bcb2-dbb70c2f9f93-scripts\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.363083 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-ovn-controller-tls-certs\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.363759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8675d591-017c-4479-bcb2-dbb70c2f9f93-combined-ca-bundle\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.381728 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9qkg\" (UniqueName: \"kubernetes.io/projected/8675d591-017c-4479-bcb2-dbb70c2f9f93-kube-api-access-m9qkg\") pod \"ovn-controller-5kwdw\" (UID: \"8675d591-017c-4479-bcb2-dbb70c2f9f93\") " pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460356 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-run\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460683 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-log\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-run\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460885 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-log\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460712 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/178e673b-646c-44a4-b368-29d3fabce0c4-scripts\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp8hk\" (UniqueName: \"kubernetes.io/projected/178e673b-646c-44a4-b368-29d3fabce0c4-kube-api-access-pp8hk\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.460979 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-lib\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.461065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-etc-ovs\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.461307 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-etc-ovs\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.461404 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/178e673b-646c-44a4-b368-29d3fabce0c4-var-lib\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.462847 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/178e673b-646c-44a4-b368-29d3fabce0c4-scripts\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.483344 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp8hk\" (UniqueName: \"kubernetes.io/projected/178e673b-646c-44a4-b368-29d3fabce0c4-kube-api-access-pp8hk\") pod \"ovn-controller-ovs-7724w\" (UID: \"178e673b-646c-44a4-b368-29d3fabce0c4\") " pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.555490 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:15 crc kubenswrapper[4726]: I1004 03:58:15.570597 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.852013 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.853576 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.858239 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.858416 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.858417 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.858668 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-srwhk" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.858780 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.883228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932480 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932555 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932597 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932640 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932816 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932885 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68fl5\" (UniqueName: \"kubernetes.io/projected/164dc913-a191-4c88-ad58-e2a08341d0a3-kube-api-access-68fl5\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:18 crc kubenswrapper[4726]: I1004 03:58:18.932940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-config\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037127 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037225 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037425 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037497 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68fl5\" (UniqueName: \"kubernetes.io/projected/164dc913-a191-4c88-ad58-e2a08341d0a3-kube-api-access-68fl5\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-config\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.037980 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.039201 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.039681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-config\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.040383 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/164dc913-a191-4c88-ad58-e2a08341d0a3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.043724 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.045805 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.048261 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.049005 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6s82l" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.050001 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.050887 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.050519 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.060464 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.062616 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.063073 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/164dc913-a191-4c88-ad58-e2a08341d0a3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.080152 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68fl5\" (UniqueName: \"kubernetes.io/projected/164dc913-a191-4c88-ad58-e2a08341d0a3-kube-api-access-68fl5\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.082303 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"164dc913-a191-4c88-ad58-e2a08341d0a3\") " pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.140454 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.140530 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d298c1d3-002d-4847-bddc-63df803a9029-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.140579 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-config\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.140818 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.140888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.141025 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2zzl\" (UniqueName: \"kubernetes.io/projected/d298c1d3-002d-4847-bddc-63df803a9029-kube-api-access-t2zzl\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.141167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.141220 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.197420 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.242842 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.242923 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2zzl\" (UniqueName: \"kubernetes.io/projected/d298c1d3-002d-4847-bddc-63df803a9029-kube-api-access-t2zzl\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.242988 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.243020 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.243077 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.243095 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d298c1d3-002d-4847-bddc-63df803a9029-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.243162 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-config\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.243243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.244430 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.245053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d298c1d3-002d-4847-bddc-63df803a9029-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.245378 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.245496 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d298c1d3-002d-4847-bddc-63df803a9029-config\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.249233 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.255953 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.257289 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d298c1d3-002d-4847-bddc-63df803a9029-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.267430 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2zzl\" (UniqueName: \"kubernetes.io/projected/d298c1d3-002d-4847-bddc-63df803a9029-kube-api-access-t2zzl\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.275752 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d298c1d3-002d-4847-bddc-63df803a9029\") " pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:19 crc kubenswrapper[4726]: I1004 03:58:19.428541 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:23 crc kubenswrapper[4726]: I1004 03:58:23.301244 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 03:58:24 crc kubenswrapper[4726]: I1004 03:58:24.320839 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 03:58:24 crc kubenswrapper[4726]: W1004 03:58:24.929927 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded4e96c4_5868_4fd7_970b_ca7c74b56b1a.slice/crio-6b2df2070b57c9b0f3e6422b3e45624a19a6ac96ef84c43069869a65993d1b08 WatchSource:0}: Error finding container 6b2df2070b57c9b0f3e6422b3e45624a19a6ac96ef84c43069869a65993d1b08: Status 404 returned error can't find the container with id 6b2df2070b57c9b0f3e6422b3e45624a19a6ac96ef84c43069869a65993d1b08 Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.942898 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.943175 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-stn5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-lxh56_openstack(fa16927f-b443-4ed4-9988-aedb52dcd8a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.945138 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" podUID="fa16927f-b443-4ed4-9988-aedb52dcd8a9" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.984469 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.984691 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jfg8p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-wjrvs_openstack(254c8ce5-3f94-47f4-8f8a-e792235c5378): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.985926 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.998098 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 03:58:24 crc kubenswrapper[4726]: E1004 03:58:24.998287 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sggsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-vtqc2_openstack(730d38e1-c902-4ff0-af36-8434cc2b94fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:58:25 crc kubenswrapper[4726]: E1004 03:58:25.000334 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" podUID="730d38e1-c902-4ff0-af36-8434cc2b94fd" Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.532513 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.619090 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7724w"] Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.647567 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 03:58:25 crc kubenswrapper[4726]: W1004 03:58:25.652611 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d1c61fb_2030_40d7_a859_7f8c30983fc1.slice/crio-a5d8ed0bb822cb8fae7ec8a90976a8f6408d1451aa2d3c2ae179d2bc7f5de0d9 WatchSource:0}: Error finding container a5d8ed0bb822cb8fae7ec8a90976a8f6408d1451aa2d3c2ae179d2bc7f5de0d9: Status 404 returned error can't find the container with id a5d8ed0bb822cb8fae7ec8a90976a8f6408d1451aa2d3c2ae179d2bc7f5de0d9 Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.743215 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw"] Oct 04 03:58:25 crc kubenswrapper[4726]: W1004 03:58:25.745946 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8675d591_017c_4479_bcb2_dbb70c2f9f93.slice/crio-26d8a350f01daa643eb43be6d277c8bad088fea1a587173668c0fcd68658e1aa WatchSource:0}: Error finding container 26d8a350f01daa643eb43be6d277c8bad088fea1a587173668c0fcd68658e1aa: Status 404 returned error can't find the container with id 26d8a350f01daa643eb43be6d277c8bad088fea1a587173668c0fcd68658e1aa Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.755415 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a","Type":"ContainerStarted","Data":"6b2df2070b57c9b0f3e6422b3e45624a19a6ac96ef84c43069869a65993d1b08"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.759800 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a4c1331-eccd-49e9-be27-242954c93434" containerID="37442543f257f649ca8cf4490fef1eaf53d22289b3aa2b48ec2bec1349e55308" exitCode=0 Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.759881 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" event={"ID":"9a4c1331-eccd-49e9-be27-242954c93434","Type":"ContainerDied","Data":"37442543f257f649ca8cf4490fef1eaf53d22289b3aa2b48ec2bec1349e55308"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.762018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7724w" event={"ID":"178e673b-646c-44a4-b368-29d3fabce0c4","Type":"ContainerStarted","Data":"ffdf7fceddf7e1cb87e07c0c11cdce1ddcd78e58e1c3f67a423d05b870f2c01e"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.781199 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706","Type":"ContainerStarted","Data":"651db194c6d6d61b8db09bea6292954960abb5565ad856ed2415d3551db87738"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.788926 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5d1c61fb-2030-40d7-a859-7f8c30983fc1","Type":"ContainerStarted","Data":"a5d8ed0bb822cb8fae7ec8a90976a8f6408d1451aa2d3c2ae179d2bc7f5de0d9"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.792090 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d148624b-4c10-4a20-b428-902273d70ac0","Type":"ContainerStarted","Data":"eb94509788b08c17a5b8c8510472eb3dfedc246307ed957917af7dd3928bba32"} Oct 04 03:58:25 crc kubenswrapper[4726]: I1004 03:58:25.837618 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 03:58:25 crc kubenswrapper[4726]: W1004 03:58:25.863242 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd298c1d3_002d_4847_bddc_63df803a9029.slice/crio-d36393c6ea0945a77d9c7db4ab9363019ca075ab9b94e11b86071189a89064b6 WatchSource:0}: Error finding container d36393c6ea0945a77d9c7db4ab9363019ca075ab9b94e11b86071189a89064b6: Status 404 returned error can't find the container with id d36393c6ea0945a77d9c7db4ab9363019ca075ab9b94e11b86071189a89064b6 Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.514312 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.594483 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stn5j\" (UniqueName: \"kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j\") pod \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.594676 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config\") pod \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\" (UID: \"fa16927f-b443-4ed4-9988-aedb52dcd8a9\") " Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.595706 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config" (OuterVolumeSpecName: "config") pod "fa16927f-b443-4ed4-9988-aedb52dcd8a9" (UID: "fa16927f-b443-4ed4-9988-aedb52dcd8a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.600482 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j" (OuterVolumeSpecName: "kube-api-access-stn5j") pod "fa16927f-b443-4ed4-9988-aedb52dcd8a9" (UID: "fa16927f-b443-4ed4-9988-aedb52dcd8a9"). InnerVolumeSpecName "kube-api-access-stn5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.667859 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.675262 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.697332 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stn5j\" (UniqueName: \"kubernetes.io/projected/fa16927f-b443-4ed4-9988-aedb52dcd8a9-kube-api-access-stn5j\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.697386 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa16927f-b443-4ed4-9988-aedb52dcd8a9-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:26 crc kubenswrapper[4726]: W1004 03:58:26.722263 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod164dc913_a191_4c88_ad58_e2a08341d0a3.slice/crio-df07bf9f7ac8667ff42a45fa73658a345581fdd41cfa887addec5ceff4c3c23f WatchSource:0}: Error finding container df07bf9f7ac8667ff42a45fa73658a345581fdd41cfa887addec5ceff4c3c23f: Status 404 returned error can't find the container with id df07bf9f7ac8667ff42a45fa73658a345581fdd41cfa887addec5ceff4c3c23f Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.798026 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sggsw\" (UniqueName: \"kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw\") pod \"730d38e1-c902-4ff0-af36-8434cc2b94fd\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.798124 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config\") pod \"730d38e1-c902-4ff0-af36-8434cc2b94fd\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.798220 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc\") pod \"730d38e1-c902-4ff0-af36-8434cc2b94fd\" (UID: \"730d38e1-c902-4ff0-af36-8434cc2b94fd\") " Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.798713 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config" (OuterVolumeSpecName: "config") pod "730d38e1-c902-4ff0-af36-8434cc2b94fd" (UID: "730d38e1-c902-4ff0-af36-8434cc2b94fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.798726 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "730d38e1-c902-4ff0-af36-8434cc2b94fd" (UID: "730d38e1-c902-4ff0-af36-8434cc2b94fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.799848 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.799843 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vtqc2" event={"ID":"730d38e1-c902-4ff0-af36-8434cc2b94fd","Type":"ContainerDied","Data":"3e76af8c25fced04341960ffacf3801fd94fbd723065edc5e003f5bbe3546cea"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.801367 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw" (OuterVolumeSpecName: "kube-api-access-sggsw") pod "730d38e1-c902-4ff0-af36-8434cc2b94fd" (UID: "730d38e1-c902-4ff0-af36-8434cc2b94fd"). InnerVolumeSpecName "kube-api-access-sggsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.803262 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" event={"ID":"9a4c1331-eccd-49e9-be27-242954c93434","Type":"ContainerStarted","Data":"b108a0d4903d41200ca2b051bd37f8ebf2ab60161e162b0cd5155a11d28ba544"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.803331 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.806078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerStarted","Data":"d3041b93f177ab188fb6aef46dddca6ee40048662fe2bdc6cdda20d8866fde54"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.807079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw" event={"ID":"8675d591-017c-4479-bcb2-dbb70c2f9f93","Type":"ContainerStarted","Data":"26d8a350f01daa643eb43be6d277c8bad088fea1a587173668c0fcd68658e1aa"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.808650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d298c1d3-002d-4847-bddc-63df803a9029","Type":"ContainerStarted","Data":"d36393c6ea0945a77d9c7db4ab9363019ca075ab9b94e11b86071189a89064b6"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.810835 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerStarted","Data":"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.812488 4726 generic.go:334] "Generic (PLEG): container finished" podID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerID="1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9" exitCode=0 Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.812557 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" event={"ID":"254c8ce5-3f94-47f4-8f8a-e792235c5378","Type":"ContainerDied","Data":"1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.813664 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.813665 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lxh56" event={"ID":"fa16927f-b443-4ed4-9988-aedb52dcd8a9","Type":"ContainerDied","Data":"b528a37c562fac0ef72c1e78f56a8807dee3aafd661381ad159ea7ae74561a7f"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.815974 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"164dc913-a191-4c88-ad58-e2a08341d0a3","Type":"ContainerStarted","Data":"df07bf9f7ac8667ff42a45fa73658a345581fdd41cfa887addec5ceff4c3c23f"} Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.821312 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" podStartSLOduration=2.893791391 podStartE2EDuration="21.821295452s" podCreationTimestamp="2025-10-04 03:58:05 +0000 UTC" firstStartedPulling="2025-10-04 03:58:06.16995132 +0000 UTC m=+1060.344574533" lastFinishedPulling="2025-10-04 03:58:25.097455381 +0000 UTC m=+1079.272078594" observedRunningTime="2025-10-04 03:58:26.817760514 +0000 UTC m=+1080.992383727" watchObservedRunningTime="2025-10-04 03:58:26.821295452 +0000 UTC m=+1080.995918665" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.903919 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sggsw\" (UniqueName: \"kubernetes.io/projected/730d38e1-c902-4ff0-af36-8434cc2b94fd-kube-api-access-sggsw\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.903972 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.903984 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/730d38e1-c902-4ff0-af36-8434cc2b94fd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.925211 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:26 crc kubenswrapper[4726]: I1004 03:58:26.931804 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxh56"] Oct 04 03:58:27 crc kubenswrapper[4726]: I1004 03:58:27.178070 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:27 crc kubenswrapper[4726]: I1004 03:58:27.183412 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vtqc2"] Oct 04 03:58:28 crc kubenswrapper[4726]: I1004 03:58:28.512765 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="730d38e1-c902-4ff0-af36-8434cc2b94fd" path="/var/lib/kubelet/pods/730d38e1-c902-4ff0-af36-8434cc2b94fd/volumes" Oct 04 03:58:28 crc kubenswrapper[4726]: I1004 03:58:28.513960 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa16927f-b443-4ed4-9988-aedb52dcd8a9" path="/var/lib/kubelet/pods/fa16927f-b443-4ed4-9988-aedb52dcd8a9/volumes" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.651184 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.710596 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.893873 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d298c1d3-002d-4847-bddc-63df803a9029","Type":"ContainerStarted","Data":"318d129ef020d93574b6db48b9605e38a2eac23f39f101b08a050e1eb4fc38ca"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.895182 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" event={"ID":"254c8ce5-3f94-47f4-8f8a-e792235c5378","Type":"ContainerStarted","Data":"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.895315 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="dnsmasq-dns" containerID="cri-o://b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5" gracePeriod=10 Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.895548 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.897875 4726 generic.go:334] "Generic (PLEG): container finished" podID="178e673b-646c-44a4-b368-29d3fabce0c4" containerID="731bd717706fa24a70d27b4c0f37cb377209b282df0988bae42664972b7657b5" exitCode=0 Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.897931 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7724w" event={"ID":"178e673b-646c-44a4-b368-29d3fabce0c4","Type":"ContainerDied","Data":"731bd717706fa24a70d27b4c0f37cb377209b282df0988bae42664972b7657b5"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.899840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5b1149f2-4d8f-4b1e-b5ec-b13f9833c706","Type":"ContainerStarted","Data":"19d87fee13d3a7fff598b138cd287f857c67bd04f5e2bb5a9d2f1dc3a77044b9"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.899981 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.904876 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5d1c61fb-2030-40d7-a859-7f8c30983fc1","Type":"ContainerStarted","Data":"c103577951bb41470de5fc25898bdba5b2c1c455d78979280fbdddf0ca82d737"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.904973 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.906183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"164dc913-a191-4c88-ad58-e2a08341d0a3","Type":"ContainerStarted","Data":"b72faf12556bd3bd93007a33b53c525ce4da7d2c191a35627e3d5c3974147a1c"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.908152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a","Type":"ContainerStarted","Data":"1bd6f58d582908eb8101b8a86732de00e1814b585026b814d23d3d61307b19bc"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.913207 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" podStartSLOduration=-9223372004.941587 podStartE2EDuration="31.913188036s" podCreationTimestamp="2025-10-04 03:58:04 +0000 UTC" firstStartedPulling="2025-10-04 03:58:05.929764982 +0000 UTC m=+1060.104388195" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:58:35.910799179 +0000 UTC m=+1090.085422392" watchObservedRunningTime="2025-10-04 03:58:35.913188036 +0000 UTC m=+1090.087811249" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.914379 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw" event={"ID":"8675d591-017c-4479-bcb2-dbb70c2f9f93","Type":"ContainerStarted","Data":"c598d3af9fe2c185cfdde2632f6f1b35957df584ebb507393979a384d1557a49"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.915403 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-5kwdw" Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.917881 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d148624b-4c10-4a20-b428-902273d70ac0","Type":"ContainerStarted","Data":"21372e099e5ad7656627a49f0d344c68f49817d0ff36fc68654802e374bbf6e9"} Oct 04 03:58:35 crc kubenswrapper[4726]: I1004 03:58:35.949138 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.36536263 podStartE2EDuration="26.949095751s" podCreationTimestamp="2025-10-04 03:58:09 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.544960585 +0000 UTC m=+1079.719583808" lastFinishedPulling="2025-10-04 03:58:33.128693706 +0000 UTC m=+1087.303316929" observedRunningTime="2025-10-04 03:58:35.934289091 +0000 UTC m=+1090.108912304" watchObservedRunningTime="2025-10-04 03:58:35.949095751 +0000 UTC m=+1090.123718964" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.000158 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.162400759 podStartE2EDuration="25.000139346s" podCreationTimestamp="2025-10-04 03:58:11 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.654528422 +0000 UTC m=+1079.829151635" lastFinishedPulling="2025-10-04 03:58:35.492266999 +0000 UTC m=+1089.666890222" observedRunningTime="2025-10-04 03:58:35.997961246 +0000 UTC m=+1090.172584459" watchObservedRunningTime="2025-10-04 03:58:36.000139346 +0000 UTC m=+1090.174762559" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.039840 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5kwdw" podStartSLOduration=12.209536882 podStartE2EDuration="21.039822066s" podCreationTimestamp="2025-10-04 03:58:15 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.75117594 +0000 UTC m=+1079.925799153" lastFinishedPulling="2025-10-04 03:58:34.581461114 +0000 UTC m=+1088.756084337" observedRunningTime="2025-10-04 03:58:36.039517547 +0000 UTC m=+1090.214140780" watchObservedRunningTime="2025-10-04 03:58:36.039822066 +0000 UTC m=+1090.214445279" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.270701 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.385720 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc\") pod \"254c8ce5-3f94-47f4-8f8a-e792235c5378\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.385766 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config\") pod \"254c8ce5-3f94-47f4-8f8a-e792235c5378\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.385931 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfg8p\" (UniqueName: \"kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p\") pod \"254c8ce5-3f94-47f4-8f8a-e792235c5378\" (UID: \"254c8ce5-3f94-47f4-8f8a-e792235c5378\") " Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.391074 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p" (OuterVolumeSpecName: "kube-api-access-jfg8p") pod "254c8ce5-3f94-47f4-8f8a-e792235c5378" (UID: "254c8ce5-3f94-47f4-8f8a-e792235c5378"). InnerVolumeSpecName "kube-api-access-jfg8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.433889 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config" (OuterVolumeSpecName: "config") pod "254c8ce5-3f94-47f4-8f8a-e792235c5378" (UID: "254c8ce5-3f94-47f4-8f8a-e792235c5378"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.434403 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "254c8ce5-3f94-47f4-8f8a-e792235c5378" (UID: "254c8ce5-3f94-47f4-8f8a-e792235c5378"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.487328 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfg8p\" (UniqueName: \"kubernetes.io/projected/254c8ce5-3f94-47f4-8f8a-e792235c5378-kube-api-access-jfg8p\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.487361 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.487372 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/254c8ce5-3f94-47f4-8f8a-e792235c5378-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.929827 4726 generic.go:334] "Generic (PLEG): container finished" podID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerID="b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5" exitCode=0 Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.929891 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.929936 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" event={"ID":"254c8ce5-3f94-47f4-8f8a-e792235c5378","Type":"ContainerDied","Data":"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5"} Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.929998 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-wjrvs" event={"ID":"254c8ce5-3f94-47f4-8f8a-e792235c5378","Type":"ContainerDied","Data":"5a6cda46ef4a079ffaeef5f546c74bca089c9d02536e6e380ea6fc0796629914"} Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.930021 4726 scope.go:117] "RemoveContainer" containerID="b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.938249 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7724w" event={"ID":"178e673b-646c-44a4-b368-29d3fabce0c4","Type":"ContainerStarted","Data":"40e7f7b61d27b0b47dfdfccdd1b5fe76f5ab6234e060372cb86f17f2f069aeec"} Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.938293 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.938308 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7724w" event={"ID":"178e673b-646c-44a4-b368-29d3fabce0c4","Type":"ContainerStarted","Data":"17fd2e4922ed298cd1788623f8a28405af67e8d66e0ef261b67e4c2badb9e1d8"} Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.939809 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.959281 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.967595 4726 scope.go:117] "RemoveContainer" containerID="1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.976033 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-wjrvs"] Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.985334 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7724w" podStartSLOduration=14.481370912 podStartE2EDuration="21.985314332s" podCreationTimestamp="2025-10-04 03:58:15 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.62380685 +0000 UTC m=+1079.798430063" lastFinishedPulling="2025-10-04 03:58:33.12775026 +0000 UTC m=+1087.302373483" observedRunningTime="2025-10-04 03:58:36.96502171 +0000 UTC m=+1091.139644933" watchObservedRunningTime="2025-10-04 03:58:36.985314332 +0000 UTC m=+1091.159937535" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.994741 4726 scope.go:117] "RemoveContainer" containerID="b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5" Oct 04 03:58:36 crc kubenswrapper[4726]: E1004 03:58:36.995232 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5\": container with ID starting with b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5 not found: ID does not exist" containerID="b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.995258 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5"} err="failed to get container status \"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5\": rpc error: code = NotFound desc = could not find container \"b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5\": container with ID starting with b03e0f351e5f6f806f7f0d128fa0487436afc2ecaea4d8c4eaeb9c6c147192b5 not found: ID does not exist" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.995278 4726 scope.go:117] "RemoveContainer" containerID="1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9" Oct 04 03:58:36 crc kubenswrapper[4726]: E1004 03:58:36.996745 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9\": container with ID starting with 1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9 not found: ID does not exist" containerID="1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9" Oct 04 03:58:36 crc kubenswrapper[4726]: I1004 03:58:36.996803 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9"} err="failed to get container status \"1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9\": rpc error: code = NotFound desc = could not find container \"1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9\": container with ID starting with 1a13b549bfffcd10f05b6b526b2b469fb30d920eae4420707ea6a2df96173df9 not found: ID does not exist" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.520496 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" path="/var/lib/kubelet/pods/254c8ce5-3f94-47f4-8f8a-e792235c5378/volumes" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.709916 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-r526s"] Oct 04 03:58:38 crc kubenswrapper[4726]: E1004 03:58:38.710524 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="init" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.710538 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="init" Oct 04 03:58:38 crc kubenswrapper[4726]: E1004 03:58:38.710560 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="dnsmasq-dns" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.710567 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="dnsmasq-dns" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.710742 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="254c8ce5-3f94-47f4-8f8a-e792235c5378" containerName="dnsmasq-dns" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.711476 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.716449 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.720850 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovn-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.720898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507a6d8a-447d-436c-918d-7e2f447705bf-config\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.720940 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovs-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.720987 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.721057 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rz5\" (UniqueName: \"kubernetes.io/projected/507a6d8a-447d-436c-918d-7e2f447705bf-kube-api-access-88rz5\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.721225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-combined-ca-bundle\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.723175 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r526s"] Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.823898 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rz5\" (UniqueName: \"kubernetes.io/projected/507a6d8a-447d-436c-918d-7e2f447705bf-kube-api-access-88rz5\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.824325 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-combined-ca-bundle\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.825409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovn-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.825448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507a6d8a-447d-436c-918d-7e2f447705bf-config\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.825486 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovs-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.825521 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.825862 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovn-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.826360 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/507a6d8a-447d-436c-918d-7e2f447705bf-ovs-rundir\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.826672 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507a6d8a-447d-436c-918d-7e2f447705bf-config\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.838002 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-combined-ca-bundle\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.843570 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xmlk6"] Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.845551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.847882 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/507a6d8a-447d-436c-918d-7e2f447705bf-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.848062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rz5\" (UniqueName: \"kubernetes.io/projected/507a6d8a-447d-436c-918d-7e2f447705bf-kube-api-access-88rz5\") pod \"ovn-controller-metrics-r526s\" (UID: \"507a6d8a-447d-436c-918d-7e2f447705bf\") " pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.849818 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.866692 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xmlk6"] Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.927548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.927720 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.927868 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gq5\" (UniqueName: \"kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.927918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:38 crc kubenswrapper[4726]: I1004 03:58:38.997446 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xmlk6"] Oct 04 03:58:38 crc kubenswrapper[4726]: E1004 03:58:38.998083 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-z7gq5 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" podUID="8e670020-716a-4a02-92ec-1e459ce86cb6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.028721 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.029374 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.029484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.029527 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gq5\" (UniqueName: \"kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.029544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.030126 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.030765 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.030825 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.031053 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.032525 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.041397 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.046224 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-r526s" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.065288 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gq5\" (UniqueName: \"kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5\") pod \"dnsmasq-dns-5bf47b49b7-xmlk6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.131307 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.131368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.131436 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4dm6\" (UniqueName: \"kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.131489 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.131546 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.233021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.233328 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.233359 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.233407 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4dm6\" (UniqueName: \"kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.233445 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.234134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.234174 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.235033 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.235699 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.255423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4dm6\" (UniqueName: \"kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6\") pod \"dnsmasq-dns-8554648995-jd995\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.343722 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.966543 4726 generic.go:334] "Generic (PLEG): container finished" podID="d148624b-4c10-4a20-b428-902273d70ac0" containerID="21372e099e5ad7656627a49f0d344c68f49817d0ff36fc68654802e374bbf6e9" exitCode=0 Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.966612 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d148624b-4c10-4a20-b428-902273d70ac0","Type":"ContainerDied","Data":"21372e099e5ad7656627a49f0d344c68f49817d0ff36fc68654802e374bbf6e9"} Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.972230 4726 generic.go:334] "Generic (PLEG): container finished" podID="ed4e96c4-5868-4fd7-970b-ca7c74b56b1a" containerID="1bd6f58d582908eb8101b8a86732de00e1814b585026b814d23d3d61307b19bc" exitCode=0 Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.972286 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a","Type":"ContainerDied","Data":"1bd6f58d582908eb8101b8a86732de00e1814b585026b814d23d3d61307b19bc"} Oct 04 03:58:39 crc kubenswrapper[4726]: I1004 03:58:39.972330 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.019283 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.116673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.161366 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7gq5\" (UniqueName: \"kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5\") pod \"8e670020-716a-4a02-92ec-1e459ce86cb6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.161950 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb\") pod \"8e670020-716a-4a02-92ec-1e459ce86cb6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.162013 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config\") pod \"8e670020-716a-4a02-92ec-1e459ce86cb6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.162053 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc\") pod \"8e670020-716a-4a02-92ec-1e459ce86cb6\" (UID: \"8e670020-716a-4a02-92ec-1e459ce86cb6\") " Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.163782 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e670020-716a-4a02-92ec-1e459ce86cb6" (UID: "8e670020-716a-4a02-92ec-1e459ce86cb6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.164127 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config" (OuterVolumeSpecName: "config") pod "8e670020-716a-4a02-92ec-1e459ce86cb6" (UID: "8e670020-716a-4a02-92ec-1e459ce86cb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.164672 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e670020-716a-4a02-92ec-1e459ce86cb6" (UID: "8e670020-716a-4a02-92ec-1e459ce86cb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.167883 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5" (OuterVolumeSpecName: "kube-api-access-z7gq5") pod "8e670020-716a-4a02-92ec-1e459ce86cb6" (UID: "8e670020-716a-4a02-92ec-1e459ce86cb6"). InnerVolumeSpecName "kube-api-access-z7gq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.263922 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.263995 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.264015 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7gq5\" (UniqueName: \"kubernetes.io/projected/8e670020-716a-4a02-92ec-1e459ce86cb6-kube-api-access-z7gq5\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.264035 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e670020-716a-4a02-92ec-1e459ce86cb6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.432724 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-r526s"] Oct 04 03:58:40 crc kubenswrapper[4726]: W1004 03:58:40.441173 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod507a6d8a_447d_436c_918d_7e2f447705bf.slice/crio-9dc36944406fae184a2b4d26457ac6c9dc59dd2544a465d0536c33f48b714956 WatchSource:0}: Error finding container 9dc36944406fae184a2b4d26457ac6c9dc59dd2544a465d0536c33f48b714956: Status 404 returned error can't find the container with id 9dc36944406fae184a2b4d26457ac6c9dc59dd2544a465d0536c33f48b714956 Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.517585 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.980911 4726 generic.go:334] "Generic (PLEG): container finished" podID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerID="741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19" exitCode=0 Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.980999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jd995" event={"ID":"f152a82c-fc94-4437-b087-6fbcfd7c2ff5","Type":"ContainerDied","Data":"741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.981038 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jd995" event={"ID":"f152a82c-fc94-4437-b087-6fbcfd7c2ff5","Type":"ContainerStarted","Data":"fa41041141e3f6aa8e5423f3fc82cbf1b4cbef7d2890de3afeba604e6ee12d9b"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.983984 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d148624b-4c10-4a20-b428-902273d70ac0","Type":"ContainerStarted","Data":"2501ac86154276494af8101e34915bb38d1c67eeea908337757a4b479624170e"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.985530 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d298c1d3-002d-4847-bddc-63df803a9029","Type":"ContainerStarted","Data":"c8019b2e30382f326519dd859725b0a8ff8675de7206c1be6fdd37d1e524c26c"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.988531 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"164dc913-a191-4c88-ad58-e2a08341d0a3","Type":"ContainerStarted","Data":"4d36f6eb5b3ec1f64bb8e27d85c8c426cc2c3b3d02c147011b665c8236aa377b"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.991569 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r526s" event={"ID":"507a6d8a-447d-436c-918d-7e2f447705bf","Type":"ContainerStarted","Data":"1e5cf52e3f461f2d26fa1ba9f10b7d67fd7097302f5aa989be3e7aab363308b3"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.991707 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-r526s" event={"ID":"507a6d8a-447d-436c-918d-7e2f447705bf","Type":"ContainerStarted","Data":"9dc36944406fae184a2b4d26457ac6c9dc59dd2544a465d0536c33f48b714956"} Oct 04 03:58:40 crc kubenswrapper[4726]: I1004 03:58:40.999093 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-xmlk6" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.000060 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ed4e96c4-5868-4fd7-970b-ca7c74b56b1a","Type":"ContainerStarted","Data":"feb8813bdc0c27f5db63e5b686f8fbe51408a53fd6a65356ea80c9d1c80f2e20"} Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.024947 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-r526s" podStartSLOduration=3.02492713 podStartE2EDuration="3.02492713s" podCreationTimestamp="2025-10-04 03:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:58:41.02132727 +0000 UTC m=+1095.195950483" watchObservedRunningTime="2025-10-04 03:58:41.02492713 +0000 UTC m=+1095.199550343" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.067236 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.05188639 podStartE2EDuration="33.067216492s" podCreationTimestamp="2025-10-04 03:58:08 +0000 UTC" firstStartedPulling="2025-10-04 03:58:24.933040554 +0000 UTC m=+1079.107663777" lastFinishedPulling="2025-10-04 03:58:33.948370636 +0000 UTC m=+1088.122993879" observedRunningTime="2025-10-04 03:58:41.061854444 +0000 UTC m=+1095.236477687" watchObservedRunningTime="2025-10-04 03:58:41.067216492 +0000 UTC m=+1095.241839715" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.093935 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.523741391 podStartE2EDuration="33.093909652s" podCreationTimestamp="2025-10-04 03:58:08 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.011328274 +0000 UTC m=+1079.185951487" lastFinishedPulling="2025-10-04 03:58:34.581496525 +0000 UTC m=+1088.756119748" observedRunningTime="2025-10-04 03:58:41.084918833 +0000 UTC m=+1095.259542056" watchObservedRunningTime="2025-10-04 03:58:41.093909652 +0000 UTC m=+1095.268532865" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.141282 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.100915762 podStartE2EDuration="23.141262065s" podCreationTimestamp="2025-10-04 03:58:18 +0000 UTC" firstStartedPulling="2025-10-04 03:58:25.917676935 +0000 UTC m=+1080.092300148" lastFinishedPulling="2025-10-04 03:58:39.958023198 +0000 UTC m=+1094.132646451" observedRunningTime="2025-10-04 03:58:41.108497577 +0000 UTC m=+1095.283120790" watchObservedRunningTime="2025-10-04 03:58:41.141262065 +0000 UTC m=+1095.315885268" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.143720 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.903958331 podStartE2EDuration="24.143710113s" podCreationTimestamp="2025-10-04 03:58:17 +0000 UTC" firstStartedPulling="2025-10-04 03:58:26.72814385 +0000 UTC m=+1080.902767063" lastFinishedPulling="2025-10-04 03:58:39.967895632 +0000 UTC m=+1094.142518845" observedRunningTime="2025-10-04 03:58:41.132212834 +0000 UTC m=+1095.306836047" watchObservedRunningTime="2025-10-04 03:58:41.143710113 +0000 UTC m=+1095.318333336" Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.167253 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xmlk6"] Oct 04 03:58:41 crc kubenswrapper[4726]: I1004 03:58:41.177088 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-xmlk6"] Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.034336 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jd995" event={"ID":"f152a82c-fc94-4437-b087-6fbcfd7c2ff5","Type":"ContainerStarted","Data":"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f"} Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.036043 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.114213 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-jd995" podStartSLOduration=3.114193142 podStartE2EDuration="3.114193142s" podCreationTimestamp="2025-10-04 03:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:58:42.086206266 +0000 UTC m=+1096.260829499" watchObservedRunningTime="2025-10-04 03:58:42.114193142 +0000 UTC m=+1096.288816355" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.126927 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.127168 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.175370 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.178068 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.204007 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.228260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.228355 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.228403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.228462 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgdm4\" (UniqueName: \"kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.228506 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.331125 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.331212 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.331269 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgdm4\" (UniqueName: \"kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.331306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.331354 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.332243 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.332813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.333446 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.334711 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.356658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgdm4\" (UniqueName: \"kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4\") pod \"dnsmasq-dns-b8fbc5445-qjv7m\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.510845 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e670020-716a-4a02-92ec-1e459ce86cb6" path="/var/lib/kubelet/pods/8e670020-716a-4a02-92ec-1e459ce86cb6/volumes" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.518597 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:42 crc kubenswrapper[4726]: I1004 03:58:42.749310 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:58:42 crc kubenswrapper[4726]: W1004 03:58:42.760236 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode415f14e_2ec6_4e9c_89dc_df30dc0d7152.slice/crio-4bda3405a87aad1783fd2ba123dcc5b8294d2bb947361391874c476605c32318 WatchSource:0}: Error finding container 4bda3405a87aad1783fd2ba123dcc5b8294d2bb947361391874c476605c32318: Status 404 returned error can't find the container with id 4bda3405a87aad1783fd2ba123dcc5b8294d2bb947361391874c476605c32318 Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.043556 4726 generic.go:334] "Generic (PLEG): container finished" podID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerID="53f0171f7f26ba1a820ef9b5fcdee5dde1bcc0b09388bfe2e43bca8df5698a2e" exitCode=0 Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.043614 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" event={"ID":"e415f14e-2ec6-4e9c-89dc-df30dc0d7152","Type":"ContainerDied","Data":"53f0171f7f26ba1a820ef9b5fcdee5dde1bcc0b09388bfe2e43bca8df5698a2e"} Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.043663 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" event={"ID":"e415f14e-2ec6-4e9c-89dc-df30dc0d7152","Type":"ContainerStarted","Data":"4bda3405a87aad1783fd2ba123dcc5b8294d2bb947361391874c476605c32318"} Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.198129 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.236946 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.244043 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.250980 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.252786 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.253016 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.253397 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-jxfcc" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.254364 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.280069 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.350068 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.350207 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-lock\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.350238 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-cache\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.350363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzxx2\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-kube-api-access-rzxx2\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.350390 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.428737 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.451921 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzxx2\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-kube-api-access-rzxx2\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.451978 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452042 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452143 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-lock\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452169 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-cache\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.452298 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.452340 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.452407 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:58:43.952381263 +0000 UTC m=+1098.127004486 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452571 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452618 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-cache\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.452883 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c4528cfc-3f8b-4850-aeaa-eabe209382bd-lock\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.466965 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.479295 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzxx2\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-kube-api-access-rzxx2\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.482541 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: I1004 03:58:43.960447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.960722 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.960768 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:43 crc kubenswrapper[4726]: E1004 03:58:43.960857 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:58:44.960827006 +0000 UTC m=+1099.135450239 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.057025 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" event={"ID":"e415f14e-2ec6-4e9c-89dc-df30dc0d7152","Type":"ContainerStarted","Data":"d8c5ab6e778403ee100c722bdc069e9a069661dfc083b54ecd9ffc4f7e0dc861"} Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.058356 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.058396 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.058565 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-jd995" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="dnsmasq-dns" containerID="cri-o://e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f" gracePeriod=10 Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.092202 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" podStartSLOduration=2.092180846 podStartE2EDuration="2.092180846s" podCreationTimestamp="2025-10-04 03:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:58:44.08364153 +0000 UTC m=+1098.258264753" watchObservedRunningTime="2025-10-04 03:58:44.092180846 +0000 UTC m=+1098.266804099" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.109093 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.118074 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.485011 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.488755 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.491648 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.491877 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.492096 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-62p5n" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.502088 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.526755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.572802 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573154 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-config\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573345 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-scripts\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74t4h\" (UniqueName: \"kubernetes.io/projected/f3ae4630-e488-43db-a57a-762c75de75aa-kube-api-access-74t4h\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.573596 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675174 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config\") pod \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675246 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc\") pod \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb\") pod \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675439 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb\") pod \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675470 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4dm6\" (UniqueName: \"kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6\") pod \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\" (UID: \"f152a82c-fc94-4437-b087-6fbcfd7c2ff5\") " Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74t4h\" (UniqueName: \"kubernetes.io/projected/f3ae4630-e488-43db-a57a-762c75de75aa-kube-api-access-74t4h\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675722 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675739 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-config\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675796 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.675831 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-scripts\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.676629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-scripts\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.677824 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.678569 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ae4630-e488-43db-a57a-762c75de75aa-config\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.684023 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.684588 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.687656 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6" (OuterVolumeSpecName: "kube-api-access-x4dm6") pod "f152a82c-fc94-4437-b087-6fbcfd7c2ff5" (UID: "f152a82c-fc94-4437-b087-6fbcfd7c2ff5"). InnerVolumeSpecName "kube-api-access-x4dm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.694866 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3ae4630-e488-43db-a57a-762c75de75aa-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.716276 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74t4h\" (UniqueName: \"kubernetes.io/projected/f3ae4630-e488-43db-a57a-762c75de75aa-kube-api-access-74t4h\") pod \"ovn-northd-0\" (UID: \"f3ae4630-e488-43db-a57a-762c75de75aa\") " pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: E1004 03:58:44.722588 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.176:45060->38.102.83.176:46161: write tcp 38.102.83.176:45060->38.102.83.176:46161: write: broken pipe Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.740875 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f152a82c-fc94-4437-b087-6fbcfd7c2ff5" (UID: "f152a82c-fc94-4437-b087-6fbcfd7c2ff5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.746477 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config" (OuterVolumeSpecName: "config") pod "f152a82c-fc94-4437-b087-6fbcfd7c2ff5" (UID: "f152a82c-fc94-4437-b087-6fbcfd7c2ff5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.757777 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f152a82c-fc94-4437-b087-6fbcfd7c2ff5" (UID: "f152a82c-fc94-4437-b087-6fbcfd7c2ff5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.762671 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f152a82c-fc94-4437-b087-6fbcfd7c2ff5" (UID: "f152a82c-fc94-4437-b087-6fbcfd7c2ff5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.776920 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.776955 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.776971 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.776985 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4dm6\" (UniqueName: \"kubernetes.io/projected/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-kube-api-access-x4dm6\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.776997 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f152a82c-fc94-4437-b087-6fbcfd7c2ff5-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.873334 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 03:58:44 crc kubenswrapper[4726]: I1004 03:58:44.979100 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:44 crc kubenswrapper[4726]: E1004 03:58:44.979324 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:44 crc kubenswrapper[4726]: E1004 03:58:44.979352 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:44 crc kubenswrapper[4726]: E1004 03:58:44.979429 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:58:46.979408128 +0000 UTC m=+1101.154031341 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.066713 4726 generic.go:334] "Generic (PLEG): container finished" podID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerID="e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f" exitCode=0 Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.066791 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-jd995" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.066811 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jd995" event={"ID":"f152a82c-fc94-4437-b087-6fbcfd7c2ff5","Type":"ContainerDied","Data":"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f"} Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.067305 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-jd995" event={"ID":"f152a82c-fc94-4437-b087-6fbcfd7c2ff5","Type":"ContainerDied","Data":"fa41041141e3f6aa8e5423f3fc82cbf1b4cbef7d2890de3afeba604e6ee12d9b"} Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.067344 4726 scope.go:117] "RemoveContainer" containerID="e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.067696 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.098305 4726 scope.go:117] "RemoveContainer" containerID="741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.117131 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.123221 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-jd995"] Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.129890 4726 scope.go:117] "RemoveContainer" containerID="e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f" Oct 04 03:58:45 crc kubenswrapper[4726]: E1004 03:58:45.130403 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f\": container with ID starting with e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f not found: ID does not exist" containerID="e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.130436 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f"} err="failed to get container status \"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f\": rpc error: code = NotFound desc = could not find container \"e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f\": container with ID starting with e349e0e31b241fa3ecd9b880ce0988a3264a6359fa513cb95f0fc322db069b2f not found: ID does not exist" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.130467 4726 scope.go:117] "RemoveContainer" containerID="741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19" Oct 04 03:58:45 crc kubenswrapper[4726]: E1004 03:58:45.130742 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19\": container with ID starting with 741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19 not found: ID does not exist" containerID="741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19" Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.130776 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19"} err="failed to get container status \"741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19\": rpc error: code = NotFound desc = could not find container \"741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19\": container with ID starting with 741a1b123e00576cc8cfe126fae08d710dae0aefa6b2dcbd722d65b5418def19 not found: ID does not exist" Oct 04 03:58:45 crc kubenswrapper[4726]: W1004 03:58:45.324339 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3ae4630_e488_43db_a57a_762c75de75aa.slice/crio-774c26bb7a15f92e2beda673d44bd1635cd23a34daee5a8953889d1c4bc36936 WatchSource:0}: Error finding container 774c26bb7a15f92e2beda673d44bd1635cd23a34daee5a8953889d1c4bc36936: Status 404 returned error can't find the container with id 774c26bb7a15f92e2beda673d44bd1635cd23a34daee5a8953889d1c4bc36936 Oct 04 03:58:45 crc kubenswrapper[4726]: I1004 03:58:45.326598 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 03:58:46 crc kubenswrapper[4726]: I1004 03:58:46.076782 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f3ae4630-e488-43db-a57a-762c75de75aa","Type":"ContainerStarted","Data":"774c26bb7a15f92e2beda673d44bd1635cd23a34daee5a8953889d1c4bc36936"} Oct 04 03:58:46 crc kubenswrapper[4726]: I1004 03:58:46.518314 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" path="/var/lib/kubelet/pods/f152a82c-fc94-4437-b087-6fbcfd7c2ff5/volumes" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.013435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:47 crc kubenswrapper[4726]: E1004 03:58:47.013642 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:47 crc kubenswrapper[4726]: E1004 03:58:47.013678 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:47 crc kubenswrapper[4726]: E1004 03:58:47.013765 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:58:51.013737911 +0000 UTC m=+1105.188361164 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.233609 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-qxprg"] Oct 04 03:58:47 crc kubenswrapper[4726]: E1004 03:58:47.233999 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="init" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.234014 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="init" Oct 04 03:58:47 crc kubenswrapper[4726]: E1004 03:58:47.234040 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="dnsmasq-dns" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.234049 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="dnsmasq-dns" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.234266 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f152a82c-fc94-4437-b087-6fbcfd7c2ff5" containerName="dnsmasq-dns" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.234880 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.238560 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.238654 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.249297 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qxprg"] Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.250524 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nbdx\" (UniqueName: \"kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420432 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420553 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420698 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.420799 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522515 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522578 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nbdx\" (UniqueName: \"kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522684 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522708 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.522797 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.523651 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.523769 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.524045 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.527406 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.528163 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.528486 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.550542 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nbdx\" (UniqueName: \"kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx\") pod \"swift-ring-rebalance-qxprg\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:47 crc kubenswrapper[4726]: I1004 03:58:47.564436 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:58:48 crc kubenswrapper[4726]: I1004 03:58:48.156479 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qxprg"] Oct 04 03:58:48 crc kubenswrapper[4726]: W1004 03:58:48.160042 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63b2f4dc_088d_47ba_aad0_e0de06d93828.slice/crio-204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376 WatchSource:0}: Error finding container 204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376: Status 404 returned error can't find the container with id 204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376 Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.111385 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qxprg" event={"ID":"63b2f4dc-088d-47ba-aad0-e0de06d93828","Type":"ContainerStarted","Data":"204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376"} Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.375332 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.375749 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.390802 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.390874 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 03:58:49 crc kubenswrapper[4726]: I1004 03:58:49.470829 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:50 crc kubenswrapper[4726]: I1004 03:58:50.169328 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 03:58:51 crc kubenswrapper[4726]: I1004 03:58:51.091756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:51 crc kubenswrapper[4726]: E1004 03:58:51.092060 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:51 crc kubenswrapper[4726]: E1004 03:58:51.092420 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:51 crc kubenswrapper[4726]: E1004 03:58:51.092523 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:58:59.092491536 +0000 UTC m=+1113.267114789 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:52 crc kubenswrapper[4726]: I1004 03:58:52.343520 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 03:58:52 crc kubenswrapper[4726]: I1004 03:58:52.414372 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d148624b-4c10-4a20-b428-902273d70ac0" containerName="galera" probeResult="failure" output=< Oct 04 03:58:52 crc kubenswrapper[4726]: wsrep_local_state_comment (Joined) differs from Synced Oct 04 03:58:52 crc kubenswrapper[4726]: > Oct 04 03:58:52 crc kubenswrapper[4726]: I1004 03:58:52.521167 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:58:52 crc kubenswrapper[4726]: I1004 03:58:52.588952 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:52 crc kubenswrapper[4726]: I1004 03:58:52.589579 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="dnsmasq-dns" containerID="cri-o://b108a0d4903d41200ca2b051bd37f8ebf2ab60161e162b0cd5155a11d28ba544" gracePeriod=10 Oct 04 03:58:53 crc kubenswrapper[4726]: I1004 03:58:53.161565 4726 generic.go:334] "Generic (PLEG): container finished" podID="9a4c1331-eccd-49e9-be27-242954c93434" containerID="b108a0d4903d41200ca2b051bd37f8ebf2ab60161e162b0cd5155a11d28ba544" exitCode=0 Oct 04 03:58:53 crc kubenswrapper[4726]: I1004 03:58:53.161632 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" event={"ID":"9a4c1331-eccd-49e9-be27-242954c93434","Type":"ContainerDied","Data":"b108a0d4903d41200ca2b051bd37f8ebf2ab60161e162b0cd5155a11d28ba544"} Oct 04 03:58:53 crc kubenswrapper[4726]: I1004 03:58:53.899206 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.045701 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc\") pod \"9a4c1331-eccd-49e9-be27-242954c93434\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.045760 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config\") pod \"9a4c1331-eccd-49e9-be27-242954c93434\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.045867 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk7hk\" (UniqueName: \"kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk\") pod \"9a4c1331-eccd-49e9-be27-242954c93434\" (UID: \"9a4c1331-eccd-49e9-be27-242954c93434\") " Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.051800 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk" (OuterVolumeSpecName: "kube-api-access-jk7hk") pod "9a4c1331-eccd-49e9-be27-242954c93434" (UID: "9a4c1331-eccd-49e9-be27-242954c93434"). InnerVolumeSpecName "kube-api-access-jk7hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.085998 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a4c1331-eccd-49e9-be27-242954c93434" (UID: "9a4c1331-eccd-49e9-be27-242954c93434"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.094233 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config" (OuterVolumeSpecName: "config") pod "9a4c1331-eccd-49e9-be27-242954c93434" (UID: "9a4c1331-eccd-49e9-be27-242954c93434"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.179786 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.180095 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4c1331-eccd-49e9-be27-242954c93434-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.180130 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk7hk\" (UniqueName: \"kubernetes.io/projected/9a4c1331-eccd-49e9-be27-242954c93434-kube-api-access-jk7hk\") on node \"crc\" DevicePath \"\"" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.190068 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" event={"ID":"9a4c1331-eccd-49e9-be27-242954c93434","Type":"ContainerDied","Data":"418d0f8fec15598b0bb61232d7869f484a3dad0c0a13919121b5a932e6101f13"} Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.190302 4726 scope.go:117] "RemoveContainer" containerID="b108a0d4903d41200ca2b051bd37f8ebf2ab60161e162b0cd5155a11d28ba544" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.190522 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ls5ds" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.212846 4726 scope.go:117] "RemoveContainer" containerID="37442543f257f649ca8cf4490fef1eaf53d22289b3aa2b48ec2bec1349e55308" Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.222834 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.228431 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ls5ds"] Oct 04 03:58:54 crc kubenswrapper[4726]: I1004 03:58:54.525372 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4c1331-eccd-49e9-be27-242954c93434" path="/var/lib/kubelet/pods/9a4c1331-eccd-49e9-be27-242954c93434/volumes" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.173779 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:58:59 crc kubenswrapper[4726]: E1004 03:58:59.173963 4726 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 03:58:59 crc kubenswrapper[4726]: E1004 03:58:59.174320 4726 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 03:58:59 crc kubenswrapper[4726]: E1004 03:58:59.174387 4726 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift podName:c4528cfc-3f8b-4850-aeaa-eabe209382bd nodeName:}" failed. No retries permitted until 2025-10-04 03:59:15.174360776 +0000 UTC m=+1129.348983999 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift") pod "swift-storage-0" (UID: "c4528cfc-3f8b-4850-aeaa-eabe209382bd") : configmap "swift-ring-files" not found Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.246841 4726 generic.go:334] "Generic (PLEG): container finished" podID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerID="d3041b93f177ab188fb6aef46dddca6ee40048662fe2bdc6cdda20d8866fde54" exitCode=0 Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.246937 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerDied","Data":"d3041b93f177ab188fb6aef46dddca6ee40048662fe2bdc6cdda20d8866fde54"} Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.248883 4726 generic.go:334] "Generic (PLEG): container finished" podID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerID="ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98" exitCode=0 Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.248930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerDied","Data":"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98"} Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.453295 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.701709 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-7rbpz"] Oct 04 03:58:59 crc kubenswrapper[4726]: E1004 03:58:59.702152 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="dnsmasq-dns" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.702170 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="dnsmasq-dns" Oct 04 03:58:59 crc kubenswrapper[4726]: E1004 03:58:59.702199 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="init" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.702208 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="init" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.702355 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4c1331-eccd-49e9-be27-242954c93434" containerName="dnsmasq-dns" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.702902 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7rbpz" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.708490 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7rbpz"] Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.786243 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqtq5\" (UniqueName: \"kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5\") pod \"keystone-db-create-7rbpz\" (UID: \"25e76b2f-db99-4f3d-8f36-68df71ff19d1\") " pod="openstack/keystone-db-create-7rbpz" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.888368 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqtq5\" (UniqueName: \"kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5\") pod \"keystone-db-create-7rbpz\" (UID: \"25e76b2f-db99-4f3d-8f36-68df71ff19d1\") " pod="openstack/keystone-db-create-7rbpz" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.908845 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqtq5\" (UniqueName: \"kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5\") pod \"keystone-db-create-7rbpz\" (UID: \"25e76b2f-db99-4f3d-8f36-68df71ff19d1\") " pod="openstack/keystone-db-create-7rbpz" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.940774 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2jttv"] Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.942495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2jttv" Oct 04 03:58:59 crc kubenswrapper[4726]: I1004 03:58:59.961808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2jttv"] Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.019706 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7rbpz" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.092236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9cts\" (UniqueName: \"kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts\") pod \"placement-db-create-2jttv\" (UID: \"3c2864ee-55ba-4488-8632-cd2e985c7ecc\") " pod="openstack/placement-db-create-2jttv" Oct 04 03:59:00 crc kubenswrapper[4726]: E1004 03:59:00.145505 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified" Oct 04 03:59:00 crc kubenswrapper[4726]: E1004 03:59:00.145702 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-northd,Image:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,Command:[/usr/bin/ovn-northd],Args:[-vfile:off -vconsole:info --n-threads=1 --ovnnb-db=ssl:ovsdbserver-nb-0.openstack.svc.cluster.local:6641 --ovnsb-db=ssl:ovsdbserver-sb-0.openstack.svc.cluster.local:6642 --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:certs,Value:n5c5h68bhf9h85h5c5h56dh64bh5dh674h668h9dh585h9dh74h8h66h5f8h5f7h8ch58fhbch577h56fh5b8h5f5h659h667hb5h65fh59bhc4hf5q,ValueFrom:nil,},EnvVar{Name:ovnnorthd-config,Value:n5c8h7ch56bh8dh8hc4h5dch9dh68h6bhb7h598h549h5dbh66fh6bh5b4h5cch5d6h55ch57fhfch588h89h5ddh5d6h65bh65bh8dhc4h67dh569q,ValueFrom:nil,},EnvVar{Name:ovnnorthd-scripts,Value:n664hd8h66ch58dh64hc9h66bhd4h558h697h67bh557hdch664h567h669h555h696h556h556h5fh5bh569hbh665h9dh4h9bh564hc8h5b7h5c4q,ValueFrom:nil,},EnvVar{Name:tls-ca-bundle.pem,Value:n55fh5b7h88h5c6h687h579h64fh58fh566h657hb4h664hbh65dh579h689h99h77h96h5f5h5c9h54dhc6hc5h5ddh68fhdfh59bh5bh5ch7dhd8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74t4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-northd-0_openstack(f3ae4630-e488-43db-a57a-762c75de75aa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.193642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9cts\" (UniqueName: \"kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts\") pod \"placement-db-create-2jttv\" (UID: \"3c2864ee-55ba-4488-8632-cd2e985c7ecc\") " pod="openstack/placement-db-create-2jttv" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.230044 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jpbkh"] Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.231664 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.239231 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9cts\" (UniqueName: \"kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts\") pod \"placement-db-create-2jttv\" (UID: \"3c2864ee-55ba-4488-8632-cd2e985c7ecc\") " pod="openstack/placement-db-create-2jttv" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.255056 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jpbkh"] Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.294851 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2jttv" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.295398 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wlz4\" (UniqueName: \"kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4\") pod \"glance-db-create-jpbkh\" (UID: \"6353dfea-176b-4703-8076-afabae252368\") " pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.397435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wlz4\" (UniqueName: \"kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4\") pod \"glance-db-create-jpbkh\" (UID: \"6353dfea-176b-4703-8076-afabae252368\") " pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.412257 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wlz4\" (UniqueName: \"kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4\") pod \"glance-db-create-jpbkh\" (UID: \"6353dfea-176b-4703-8076-afabae252368\") " pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:00 crc kubenswrapper[4726]: I1004 03:59:00.580458 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:01 crc kubenswrapper[4726]: E1004 03:59:01.895918 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-northd-0" podUID="f3ae4630-e488-43db-a57a-762c75de75aa" Oct 04 03:59:01 crc kubenswrapper[4726]: I1004 03:59:01.981644 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2jttv"] Oct 04 03:59:01 crc kubenswrapper[4726]: W1004 03:59:01.982722 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c2864ee_55ba_4488_8632_cd2e985c7ecc.slice/crio-b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a WatchSource:0}: Error finding container b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a: Status 404 returned error can't find the container with id b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.060092 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-7rbpz"] Oct 04 03:59:02 crc kubenswrapper[4726]: W1004 03:59:02.062199 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25e76b2f_db99_4f3d_8f36_68df71ff19d1.slice/crio-d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a WatchSource:0}: Error finding container d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a: Status 404 returned error can't find the container with id d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.066898 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jpbkh"] Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.278886 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerStarted","Data":"6371c43e8fec9ebcb3a9d3818bf3e23c8d70a5559ddcda6ea9235314e6623f77"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.279421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.282135 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f3ae4630-e488-43db-a57a-762c75de75aa","Type":"ContainerStarted","Data":"dc3f63f332a151fa49a9094e5b4bd4ddcb90743c8be5f6f701c7146c58f3dac5"} Oct 04 03:59:02 crc kubenswrapper[4726]: E1004 03:59:02.283773 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="f3ae4630-e488-43db-a57a-762c75de75aa" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.283922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qxprg" event={"ID":"63b2f4dc-088d-47ba-aad0-e0de06d93828","Type":"ContainerStarted","Data":"ac5bfd054f2fdab30cc42cde46692f42a0538ac31eec817be9d6c8ff9af1525f"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.286068 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerStarted","Data":"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.286462 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.287729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2jttv" event={"ID":"3c2864ee-55ba-4488-8632-cd2e985c7ecc","Type":"ContainerStarted","Data":"0bb9e6de62e1c70b8401b93e237221c882313b93df7a747834e5798b35fb282b"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.287785 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2jttv" event={"ID":"3c2864ee-55ba-4488-8632-cd2e985c7ecc","Type":"ContainerStarted","Data":"b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.289617 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7rbpz" event={"ID":"25e76b2f-db99-4f3d-8f36-68df71ff19d1","Type":"ContainerStarted","Data":"a640bb89d686ebbab0a02db0033f70a8db505b6c165d09ed4fa3753a59c93a82"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.289661 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7rbpz" event={"ID":"25e76b2f-db99-4f3d-8f36-68df71ff19d1","Type":"ContainerStarted","Data":"d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.291300 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jpbkh" event={"ID":"6353dfea-176b-4703-8076-afabae252368","Type":"ContainerStarted","Data":"5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.291335 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jpbkh" event={"ID":"6353dfea-176b-4703-8076-afabae252368","Type":"ContainerStarted","Data":"d569d8a66d70e088b3ab7ec4b53fe2f2836e0ea24cd3f3b82a8ab95a584cee5b"} Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.314390 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.823905533 podStartE2EDuration="57.314371502s" podCreationTimestamp="2025-10-04 03:58:05 +0000 UTC" firstStartedPulling="2025-10-04 03:58:10.609408499 +0000 UTC m=+1064.784031722" lastFinishedPulling="2025-10-04 03:58:25.099874478 +0000 UTC m=+1079.274497691" observedRunningTime="2025-10-04 03:59:02.312132361 +0000 UTC m=+1116.486755594" watchObservedRunningTime="2025-10-04 03:59:02.314371502 +0000 UTC m=+1116.488994715" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.337680 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-7rbpz" podStartSLOduration=3.337661878 podStartE2EDuration="3.337661878s" podCreationTimestamp="2025-10-04 03:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:02.335493908 +0000 UTC m=+1116.510117121" watchObservedRunningTime="2025-10-04 03:59:02.337661878 +0000 UTC m=+1116.512285091" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.358578 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.256173259 podStartE2EDuration="57.358562078s" podCreationTimestamp="2025-10-04 03:58:05 +0000 UTC" firstStartedPulling="2025-10-04 03:58:09.994917948 +0000 UTC m=+1064.169541161" lastFinishedPulling="2025-10-04 03:58:25.097306767 +0000 UTC m=+1079.271929980" observedRunningTime="2025-10-04 03:59:02.357660317 +0000 UTC m=+1116.532283530" watchObservedRunningTime="2025-10-04 03:59:02.358562078 +0000 UTC m=+1116.533185291" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.376516 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-2jttv" podStartSLOduration=3.376498021 podStartE2EDuration="3.376498021s" podCreationTimestamp="2025-10-04 03:58:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:02.372448517 +0000 UTC m=+1116.547071750" watchObservedRunningTime="2025-10-04 03:59:02.376498021 +0000 UTC m=+1116.551121234" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.397476 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-jpbkh" podStartSLOduration=2.397457042 podStartE2EDuration="2.397457042s" podCreationTimestamp="2025-10-04 03:59:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:02.393660725 +0000 UTC m=+1116.568283938" watchObservedRunningTime="2025-10-04 03:59:02.397457042 +0000 UTC m=+1116.572080255" Oct 04 03:59:02 crc kubenswrapper[4726]: I1004 03:59:02.440587 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-qxprg" podStartSLOduration=2.078749068 podStartE2EDuration="15.440574903s" podCreationTimestamp="2025-10-04 03:58:47 +0000 UTC" firstStartedPulling="2025-10-04 03:58:48.161932644 +0000 UTC m=+1102.336555857" lastFinishedPulling="2025-10-04 03:59:01.523758479 +0000 UTC m=+1115.698381692" observedRunningTime="2025-10-04 03:59:02.434620007 +0000 UTC m=+1116.609243240" watchObservedRunningTime="2025-10-04 03:59:02.440574903 +0000 UTC m=+1116.615198116" Oct 04 03:59:02 crc kubenswrapper[4726]: E1004 03:59:02.756488 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6353dfea_176b_4703_8076_afabae252368.slice/crio-5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6353dfea_176b_4703_8076_afabae252368.slice/crio-conmon-5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.305384 4726 generic.go:334] "Generic (PLEG): container finished" podID="25e76b2f-db99-4f3d-8f36-68df71ff19d1" containerID="a640bb89d686ebbab0a02db0033f70a8db505b6c165d09ed4fa3753a59c93a82" exitCode=0 Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.305986 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7rbpz" event={"ID":"25e76b2f-db99-4f3d-8f36-68df71ff19d1","Type":"ContainerDied","Data":"a640bb89d686ebbab0a02db0033f70a8db505b6c165d09ed4fa3753a59c93a82"} Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.308656 4726 generic.go:334] "Generic (PLEG): container finished" podID="6353dfea-176b-4703-8076-afabae252368" containerID="5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1" exitCode=0 Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.308701 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jpbkh" event={"ID":"6353dfea-176b-4703-8076-afabae252368","Type":"ContainerDied","Data":"5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1"} Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.310889 4726 generic.go:334] "Generic (PLEG): container finished" podID="3c2864ee-55ba-4488-8632-cd2e985c7ecc" containerID="0bb9e6de62e1c70b8401b93e237221c882313b93df7a747834e5798b35fb282b" exitCode=0 Oct 04 03:59:03 crc kubenswrapper[4726]: I1004 03:59:03.311862 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2jttv" event={"ID":"3c2864ee-55ba-4488-8632-cd2e985c7ecc","Type":"ContainerDied","Data":"0bb9e6de62e1c70b8401b93e237221c882313b93df7a747834e5798b35fb282b"} Oct 04 03:59:03 crc kubenswrapper[4726]: E1004 03:59:03.314488 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="f3ae4630-e488-43db-a57a-762c75de75aa" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.188772 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.188836 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.658439 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.770590 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7rbpz" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.775886 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2jttv" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.777659 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wlz4\" (UniqueName: \"kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4\") pod \"6353dfea-176b-4703-8076-afabae252368\" (UID: \"6353dfea-176b-4703-8076-afabae252368\") " Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.796132 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4" (OuterVolumeSpecName: "kube-api-access-2wlz4") pod "6353dfea-176b-4703-8076-afabae252368" (UID: "6353dfea-176b-4703-8076-afabae252368"). InnerVolumeSpecName "kube-api-access-2wlz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.879259 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9cts\" (UniqueName: \"kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts\") pod \"3c2864ee-55ba-4488-8632-cd2e985c7ecc\" (UID: \"3c2864ee-55ba-4488-8632-cd2e985c7ecc\") " Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.879308 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqtq5\" (UniqueName: \"kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5\") pod \"25e76b2f-db99-4f3d-8f36-68df71ff19d1\" (UID: \"25e76b2f-db99-4f3d-8f36-68df71ff19d1\") " Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.879622 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wlz4\" (UniqueName: \"kubernetes.io/projected/6353dfea-176b-4703-8076-afabae252368-kube-api-access-2wlz4\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.884383 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts" (OuterVolumeSpecName: "kube-api-access-s9cts") pod "3c2864ee-55ba-4488-8632-cd2e985c7ecc" (UID: "3c2864ee-55ba-4488-8632-cd2e985c7ecc"). InnerVolumeSpecName "kube-api-access-s9cts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.884439 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5" (OuterVolumeSpecName: "kube-api-access-tqtq5") pod "25e76b2f-db99-4f3d-8f36-68df71ff19d1" (UID: "25e76b2f-db99-4f3d-8f36-68df71ff19d1"). InnerVolumeSpecName "kube-api-access-tqtq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.981228 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqtq5\" (UniqueName: \"kubernetes.io/projected/25e76b2f-db99-4f3d-8f36-68df71ff19d1-kube-api-access-tqtq5\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:04 crc kubenswrapper[4726]: I1004 03:59:04.981267 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9cts\" (UniqueName: \"kubernetes.io/projected/3c2864ee-55ba-4488-8632-cd2e985c7ecc-kube-api-access-s9cts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.331349 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2jttv" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.331304 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2jttv" event={"ID":"3c2864ee-55ba-4488-8632-cd2e985c7ecc","Type":"ContainerDied","Data":"b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a"} Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.331941 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b57759b68db101cc29dfd99c17c593abdbb32d593bb14f2a2db8807283a1fb5a" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.332738 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-7rbpz" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.332739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-7rbpz" event={"ID":"25e76b2f-db99-4f3d-8f36-68df71ff19d1","Type":"ContainerDied","Data":"d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a"} Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.332815 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5cd7ceea996110d3d889accb681ac242560494a267db025064e1f288d9f579a" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.333910 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jpbkh" event={"ID":"6353dfea-176b-4703-8076-afabae252368","Type":"ContainerDied","Data":"d569d8a66d70e088b3ab7ec4b53fe2f2836e0ea24cd3f3b82a8ab95a584cee5b"} Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.333941 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d569d8a66d70e088b3ab7ec4b53fe2f2836e0ea24cd3f3b82a8ab95a584cee5b" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.333995 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jpbkh" Oct 04 03:59:05 crc kubenswrapper[4726]: I1004 03:59:05.621530 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5kwdw" podUID="8675d591-017c-4479-bcb2-dbb70c2f9f93" containerName="ovn-controller" probeResult="failure" output=< Oct 04 03:59:05 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 03:59:05 crc kubenswrapper[4726]: > Oct 04 03:59:08 crc kubenswrapper[4726]: I1004 03:59:08.379909 4726 generic.go:334] "Generic (PLEG): container finished" podID="63b2f4dc-088d-47ba-aad0-e0de06d93828" containerID="ac5bfd054f2fdab30cc42cde46692f42a0538ac31eec817be9d6c8ff9af1525f" exitCode=0 Oct 04 03:59:08 crc kubenswrapper[4726]: I1004 03:59:08.380406 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qxprg" event={"ID":"63b2f4dc-088d-47ba-aad0-e0de06d93828","Type":"ContainerDied","Data":"ac5bfd054f2fdab30cc42cde46692f42a0538ac31eec817be9d6c8ff9af1525f"} Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.802141 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.865876 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.865971 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.866102 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nbdx\" (UniqueName: \"kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.866177 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.866268 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.866311 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.866349 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices\") pod \"63b2f4dc-088d-47ba-aad0-e0de06d93828\" (UID: \"63b2f4dc-088d-47ba-aad0-e0de06d93828\") " Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.867298 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.867321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.873654 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx" (OuterVolumeSpecName: "kube-api-access-9nbdx") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "kube-api-access-9nbdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.892182 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.899792 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.902445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.912477 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts" (OuterVolumeSpecName: "scripts") pod "63b2f4dc-088d-47ba-aad0-e0de06d93828" (UID: "63b2f4dc-088d-47ba-aad0-e0de06d93828"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968645 4726 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968683 4726 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968698 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nbdx\" (UniqueName: \"kubernetes.io/projected/63b2f4dc-088d-47ba-aad0-e0de06d93828-kube-api-access-9nbdx\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968712 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63b2f4dc-088d-47ba-aad0-e0de06d93828-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968724 4726 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63b2f4dc-088d-47ba-aad0-e0de06d93828-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968734 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:09 crc kubenswrapper[4726]: I1004 03:59:09.968745 4726 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63b2f4dc-088d-47ba-aad0-e0de06d93828-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.352776 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-8af8-account-create-gvlrl"] Oct 04 03:59:10 crc kubenswrapper[4726]: E1004 03:59:10.353541 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e76b2f-db99-4f3d-8f36-68df71ff19d1" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353565 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e76b2f-db99-4f3d-8f36-68df71ff19d1" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: E1004 03:59:10.353588 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b2f4dc-088d-47ba-aad0-e0de06d93828" containerName="swift-ring-rebalance" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353597 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b2f4dc-088d-47ba-aad0-e0de06d93828" containerName="swift-ring-rebalance" Oct 04 03:59:10 crc kubenswrapper[4726]: E1004 03:59:10.353616 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2864ee-55ba-4488-8632-cd2e985c7ecc" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353624 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2864ee-55ba-4488-8632-cd2e985c7ecc" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: E1004 03:59:10.353643 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6353dfea-176b-4703-8076-afabae252368" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353654 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6353dfea-176b-4703-8076-afabae252368" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353864 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6353dfea-176b-4703-8076-afabae252368" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353885 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2864ee-55ba-4488-8632-cd2e985c7ecc" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353897 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b2f4dc-088d-47ba-aad0-e0de06d93828" containerName="swift-ring-rebalance" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.353914 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e76b2f-db99-4f3d-8f36-68df71ff19d1" containerName="mariadb-database-create" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.354599 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.356555 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.364925 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8af8-account-create-gvlrl"] Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.405764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qxprg" event={"ID":"63b2f4dc-088d-47ba-aad0-e0de06d93828","Type":"ContainerDied","Data":"204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376"} Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.405845 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qxprg" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.405840 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="204eb942956d406400df3dcbddec91e70365fdfca19474a46f0171e2eb2ea376" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.478036 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bvm9\" (UniqueName: \"kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9\") pod \"glance-8af8-account-create-gvlrl\" (UID: \"6082cf2a-50dc-4e4e-8908-12b573c30b02\") " pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.580146 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bvm9\" (UniqueName: \"kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9\") pod \"glance-8af8-account-create-gvlrl\" (UID: \"6082cf2a-50dc-4e4e-8908-12b573c30b02\") " pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.596732 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5kwdw" podUID="8675d591-017c-4479-bcb2-dbb70c2f9f93" containerName="ovn-controller" probeResult="failure" output=< Oct 04 03:59:10 crc kubenswrapper[4726]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 03:59:10 crc kubenswrapper[4726]: > Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.608989 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bvm9\" (UniqueName: \"kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9\") pod \"glance-8af8-account-create-gvlrl\" (UID: \"6082cf2a-50dc-4e4e-8908-12b573c30b02\") " pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.610678 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.624192 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7724w" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.679765 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.863996 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5kwdw-config-dt56m"] Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.866042 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.869164 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.877195 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw-config-dt56m"] Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987476 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987531 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987571 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987594 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:10 crc kubenswrapper[4726]: I1004 03:59:10.987916 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4dv\" (UniqueName: \"kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089616 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089664 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089705 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089742 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.089761 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4dv\" (UniqueName: \"kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.090148 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.090174 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.090148 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.090668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.093014 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.107044 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4dv\" (UniqueName: \"kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv\") pod \"ovn-controller-5kwdw-config-dt56m\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.162079 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8af8-account-create-gvlrl"] Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.187057 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.417700 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8af8-account-create-gvlrl" event={"ID":"6082cf2a-50dc-4e4e-8908-12b573c30b02","Type":"ContainerStarted","Data":"9d088da5d54b7350c1ee13c6d593be12534721ecf5fd38fb74c8a98fb39f1ac2"} Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.418138 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8af8-account-create-gvlrl" event={"ID":"6082cf2a-50dc-4e4e-8908-12b573c30b02","Type":"ContainerStarted","Data":"d8b143b517cfe705161b50eb4aa44e62592ddaddfa4274b38f1682d6316baca3"} Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.432971 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-8af8-account-create-gvlrl" podStartSLOduration=1.432954232 podStartE2EDuration="1.432954232s" podCreationTimestamp="2025-10-04 03:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:11.432778688 +0000 UTC m=+1125.607401921" watchObservedRunningTime="2025-10-04 03:59:11.432954232 +0000 UTC m=+1125.607577435" Oct 04 03:59:11 crc kubenswrapper[4726]: I1004 03:59:11.651808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw-config-dt56m"] Oct 04 03:59:11 crc kubenswrapper[4726]: W1004 03:59:11.653036 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8733fe38_b8a1_465d_9050_d8ed5aa700b9.slice/crio-fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab WatchSource:0}: Error finding container fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab: Status 404 returned error can't find the container with id fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab Oct 04 03:59:12 crc kubenswrapper[4726]: I1004 03:59:12.431999 4726 generic.go:334] "Generic (PLEG): container finished" podID="8733fe38-b8a1-465d-9050-d8ed5aa700b9" containerID="70373a7d6931b053b6b9de4aabd4b8ca0ccbfea6b4803cea33c1113fe542dbb7" exitCode=0 Oct 04 03:59:12 crc kubenswrapper[4726]: I1004 03:59:12.432224 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-dt56m" event={"ID":"8733fe38-b8a1-465d-9050-d8ed5aa700b9","Type":"ContainerDied","Data":"70373a7d6931b053b6b9de4aabd4b8ca0ccbfea6b4803cea33c1113fe542dbb7"} Oct 04 03:59:12 crc kubenswrapper[4726]: I1004 03:59:12.432494 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-dt56m" event={"ID":"8733fe38-b8a1-465d-9050-d8ed5aa700b9","Type":"ContainerStarted","Data":"fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab"} Oct 04 03:59:12 crc kubenswrapper[4726]: I1004 03:59:12.437295 4726 generic.go:334] "Generic (PLEG): container finished" podID="6082cf2a-50dc-4e4e-8908-12b573c30b02" containerID="9d088da5d54b7350c1ee13c6d593be12534721ecf5fd38fb74c8a98fb39f1ac2" exitCode=0 Oct 04 03:59:12 crc kubenswrapper[4726]: I1004 03:59:12.437347 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8af8-account-create-gvlrl" event={"ID":"6082cf2a-50dc-4e4e-8908-12b573c30b02","Type":"ContainerDied","Data":"9d088da5d54b7350c1ee13c6d593be12534721ecf5fd38fb74c8a98fb39f1ac2"} Oct 04 03:59:13 crc kubenswrapper[4726]: I1004 03:59:13.923239 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:13 crc kubenswrapper[4726]: I1004 03:59:13.938221 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.047951 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048174 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048216 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048261 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bvm9\" (UniqueName: \"kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9\") pod \"6082cf2a-50dc-4e4e-8908-12b573c30b02\" (UID: \"6082cf2a-50dc-4e4e-8908-12b573c30b02\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048340 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4dv\" (UniqueName: \"kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048445 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048518 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn\") pod \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\" (UID: \"8733fe38-b8a1-465d-9050-d8ed5aa700b9\") " Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048728 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048829 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.048878 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run" (OuterVolumeSpecName: "var-run") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.049259 4726 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.049279 4726 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.049293 4726 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8733fe38-b8a1-465d-9050-d8ed5aa700b9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.050785 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.051503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts" (OuterVolumeSpecName: "scripts") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.055712 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9" (OuterVolumeSpecName: "kube-api-access-4bvm9") pod "6082cf2a-50dc-4e4e-8908-12b573c30b02" (UID: "6082cf2a-50dc-4e4e-8908-12b573c30b02"). InnerVolumeSpecName "kube-api-access-4bvm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.056429 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv" (OuterVolumeSpecName: "kube-api-access-mc4dv") pod "8733fe38-b8a1-465d-9050-d8ed5aa700b9" (UID: "8733fe38-b8a1-465d-9050-d8ed5aa700b9"). InnerVolumeSpecName "kube-api-access-mc4dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.151065 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4dv\" (UniqueName: \"kubernetes.io/projected/8733fe38-b8a1-465d-9050-d8ed5aa700b9-kube-api-access-mc4dv\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.151124 4726 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.151137 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8733fe38-b8a1-465d-9050-d8ed5aa700b9-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.151147 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bvm9\" (UniqueName: \"kubernetes.io/projected/6082cf2a-50dc-4e4e-8908-12b573c30b02-kube-api-access-4bvm9\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.469636 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8af8-account-create-gvlrl" event={"ID":"6082cf2a-50dc-4e4e-8908-12b573c30b02","Type":"ContainerDied","Data":"d8b143b517cfe705161b50eb4aa44e62592ddaddfa4274b38f1682d6316baca3"} Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.469676 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8b143b517cfe705161b50eb4aa44e62592ddaddfa4274b38f1682d6316baca3" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.469722 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8af8-account-create-gvlrl" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.476489 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-dt56m" event={"ID":"8733fe38-b8a1-465d-9050-d8ed5aa700b9","Type":"ContainerDied","Data":"fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab"} Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.476540 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fde075d9fa26b73ae2fa4835424d52e8d683d88323ebde8961ed1d4a080c82ab" Oct 04 03:59:14 crc kubenswrapper[4726]: I1004 03:59:14.476556 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-dt56m" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.087580 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5kwdw-config-dt56m"] Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.095969 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5kwdw-config-dt56m"] Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.187846 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5kwdw-config-xngdk"] Oct 04 03:59:15 crc kubenswrapper[4726]: E1004 03:59:15.188742 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8733fe38-b8a1-465d-9050-d8ed5aa700b9" containerName="ovn-config" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.188776 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8733fe38-b8a1-465d-9050-d8ed5aa700b9" containerName="ovn-config" Oct 04 03:59:15 crc kubenswrapper[4726]: E1004 03:59:15.188828 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6082cf2a-50dc-4e4e-8908-12b573c30b02" containerName="mariadb-account-create" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.188841 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6082cf2a-50dc-4e4e-8908-12b573c30b02" containerName="mariadb-account-create" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.189146 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6082cf2a-50dc-4e4e-8908-12b573c30b02" containerName="mariadb-account-create" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.189182 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8733fe38-b8a1-465d-9050-d8ed5aa700b9" containerName="ovn-config" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.189960 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.195780 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.197194 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw-config-xngdk"] Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.270965 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271009 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271065 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271119 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zrkl\" (UniqueName: \"kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271349 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.271436 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.277134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c4528cfc-3f8b-4850-aeaa-eabe209382bd-etc-swift\") pod \"swift-storage-0\" (UID: \"c4528cfc-3f8b-4850-aeaa-eabe209382bd\") " pod="openstack/swift-storage-0" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.368128 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376167 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376392 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zrkl\" (UniqueName: \"kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376453 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376591 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.376901 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.378000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.398532 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zrkl\" (UniqueName: \"kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl\") pod \"ovn-controller-5kwdw-config-xngdk\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.486990 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4kwmz"] Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.489826 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.505944 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8sqhw" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.506181 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.521499 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.521870 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4kwmz"] Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.582643 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.582707 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jc4s\" (UniqueName: \"kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.582795 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.582854 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.683878 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.683959 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.684016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.684034 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jc4s\" (UniqueName: \"kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.692356 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.692711 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.692923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.699033 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-5kwdw" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.701581 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jc4s\" (UniqueName: \"kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s\") pod \"glance-db-sync-4kwmz\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:15 crc kubenswrapper[4726]: I1004 03:59:15.848061 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.058342 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 03:59:16 crc kubenswrapper[4726]: W1004 03:59:16.062062 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4528cfc_3f8b_4850_aeaa_eabe209382bd.slice/crio-b6e5da99f2d62748b06e40269e49f9cf41d86021568bdb07566ee43334563707 WatchSource:0}: Error finding container b6e5da99f2d62748b06e40269e49f9cf41d86021568bdb07566ee43334563707: Status 404 returned error can't find the container with id b6e5da99f2d62748b06e40269e49f9cf41d86021568bdb07566ee43334563707 Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.092749 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5kwdw-config-xngdk"] Oct 04 03:59:16 crc kubenswrapper[4726]: W1004 03:59:16.138141 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd02c4f5_067a_47c7_9ab5_47316427397b.slice/crio-518bb0483190be4530f8ef976f597ff0497ebcc85a0fa9bc11fdf5e3c172b03c WatchSource:0}: Error finding container 518bb0483190be4530f8ef976f597ff0497ebcc85a0fa9bc11fdf5e3c172b03c: Status 404 returned error can't find the container with id 518bb0483190be4530f8ef976f597ff0497ebcc85a0fa9bc11fdf5e3c172b03c Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.334818 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4kwmz"] Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.495348 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.525556 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8733fe38-b8a1-465d-9050-d8ed5aa700b9" path="/var/lib/kubelet/pods/8733fe38-b8a1-465d-9050-d8ed5aa700b9/volumes" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.578836 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"b6e5da99f2d62748b06e40269e49f9cf41d86021568bdb07566ee43334563707"} Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.582180 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4kwmz" event={"ID":"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a","Type":"ContainerStarted","Data":"c822889ec6a2c293f0cd8cb77a37d926ef9116377854f91efbcfe0b99aa160d6"} Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.584672 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"f3ae4630-e488-43db-a57a-762c75de75aa","Type":"ContainerStarted","Data":"8b57022ade6e8516da9b36bfca421b938f787ee16e503c8585cb40172d73a783"} Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.584881 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.586703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-xngdk" event={"ID":"bd02c4f5-067a-47c7-9ab5-47316427397b","Type":"ContainerStarted","Data":"ae17c06d117565d3fc29d4864e8b4a25223a3a1011f13f7a5aa85c4889cd6d9c"} Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.586737 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-xngdk" event={"ID":"bd02c4f5-067a-47c7-9ab5-47316427397b","Type":"ContainerStarted","Data":"518bb0483190be4530f8ef976f597ff0497ebcc85a0fa9bc11fdf5e3c172b03c"} Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.716428 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5kwdw-config-xngdk" podStartSLOduration=1.7164068989999999 podStartE2EDuration="1.716406899s" podCreationTimestamp="2025-10-04 03:59:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:16.697279249 +0000 UTC m=+1130.871902462" watchObservedRunningTime="2025-10-04 03:59:16.716406899 +0000 UTC m=+1130.891030102" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.723745 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.907868312 podStartE2EDuration="32.723727117s" podCreationTimestamp="2025-10-04 03:58:44 +0000 UTC" firstStartedPulling="2025-10-04 03:58:45.328089245 +0000 UTC m=+1099.502712508" lastFinishedPulling="2025-10-04 03:59:16.1439481 +0000 UTC m=+1130.318571313" observedRunningTime="2025-10-04 03:59:16.719018349 +0000 UTC m=+1130.893641562" watchObservedRunningTime="2025-10-04 03:59:16.723727117 +0000 UTC m=+1130.898350330" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.801253 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.824542 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rlls9"] Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.881900 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rlls9"] Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.881995 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.909678 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4gh7\" (UniqueName: \"kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7\") pod \"cinder-db-create-rlls9\" (UID: \"cdee62dd-66dd-42da-840c-520048edfd3e\") " pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.933435 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lfqcg"] Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.935186 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:16 crc kubenswrapper[4726]: I1004 03:59:16.943624 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lfqcg"] Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.011907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfgm8\" (UniqueName: \"kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8\") pod \"barbican-db-create-lfqcg\" (UID: \"a09631ae-f0ad-46af-b441-76ce38840f42\") " pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.011993 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4gh7\" (UniqueName: \"kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7\") pod \"cinder-db-create-rlls9\" (UID: \"cdee62dd-66dd-42da-840c-520048edfd3e\") " pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.031501 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4gh7\" (UniqueName: \"kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7\") pod \"cinder-db-create-rlls9\" (UID: \"cdee62dd-66dd-42da-840c-520048edfd3e\") " pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.113082 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfgm8\" (UniqueName: \"kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8\") pod \"barbican-db-create-lfqcg\" (UID: \"a09631ae-f0ad-46af-b441-76ce38840f42\") " pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.130232 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfgm8\" (UniqueName: \"kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8\") pod \"barbican-db-create-lfqcg\" (UID: \"a09631ae-f0ad-46af-b441-76ce38840f42\") " pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.211683 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.221950 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xszts"] Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.222982 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xszts" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.253520 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.255013 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xszts"] Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.315798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c2xp\" (UniqueName: \"kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp\") pod \"neutron-db-create-xszts\" (UID: \"e54464f6-5188-4631-a0d3-609ce5b2584c\") " pod="openstack/neutron-db-create-xszts" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.417073 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c2xp\" (UniqueName: \"kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp\") pod \"neutron-db-create-xszts\" (UID: \"e54464f6-5188-4631-a0d3-609ce5b2584c\") " pod="openstack/neutron-db-create-xszts" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.454237 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c2xp\" (UniqueName: \"kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp\") pod \"neutron-db-create-xszts\" (UID: \"e54464f6-5188-4631-a0d3-609ce5b2584c\") " pod="openstack/neutron-db-create-xszts" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.553443 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xszts" Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.602094 4726 generic.go:334] "Generic (PLEG): container finished" podID="bd02c4f5-067a-47c7-9ab5-47316427397b" containerID="ae17c06d117565d3fc29d4864e8b4a25223a3a1011f13f7a5aa85c4889cd6d9c" exitCode=0 Oct 04 03:59:17 crc kubenswrapper[4726]: I1004 03:59:17.602645 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5kwdw-config-xngdk" event={"ID":"bd02c4f5-067a-47c7-9ab5-47316427397b","Type":"ContainerDied","Data":"ae17c06d117565d3fc29d4864e8b4a25223a3a1011f13f7a5aa85c4889cd6d9c"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.143906 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rlls9"] Oct 04 03:59:18 crc kubenswrapper[4726]: W1004 03:59:18.152087 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdee62dd_66dd_42da_840c_520048edfd3e.slice/crio-967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208 WatchSource:0}: Error finding container 967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208: Status 404 returned error can't find the container with id 967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208 Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.153372 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lfqcg"] Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.257808 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xszts"] Oct 04 03:59:18 crc kubenswrapper[4726]: W1004 03:59:18.282024 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode54464f6_5188_4631_a0d3_609ce5b2584c.slice/crio-e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b WatchSource:0}: Error finding container e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b: Status 404 returned error can't find the container with id e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.613630 4726 generic.go:334] "Generic (PLEG): container finished" podID="cdee62dd-66dd-42da-840c-520048edfd3e" containerID="5fa8851085155a12f785310916c594cfef41618ba305c4a842fb328eeb4cc89a" exitCode=0 Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.613771 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlls9" event={"ID":"cdee62dd-66dd-42da-840c-520048edfd3e","Type":"ContainerDied","Data":"5fa8851085155a12f785310916c594cfef41618ba305c4a842fb328eeb4cc89a"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.614410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlls9" event={"ID":"cdee62dd-66dd-42da-840c-520048edfd3e","Type":"ContainerStarted","Data":"967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.617960 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"b032e5a1996db9d1d99510766c0bdf0f8bb16efba3166487bbff1bcd31e4ab88"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.617988 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"2d22e58a24d06cb39e4dabf798bf725f0dceedf8cb01789442cd9728588e94f4"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.617999 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"83a7ffdde3d97ddcb80071d811bb65c6a5fc72485d2a6b8fc895be58630687aa"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.620728 4726 generic.go:334] "Generic (PLEG): container finished" podID="a09631ae-f0ad-46af-b441-76ce38840f42" containerID="0e40940d5d00393fab8ee50689d4119e0c30bb714227eba711a8375f701c7c9e" exitCode=0 Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.620838 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lfqcg" event={"ID":"a09631ae-f0ad-46af-b441-76ce38840f42","Type":"ContainerDied","Data":"0e40940d5d00393fab8ee50689d4119e0c30bb714227eba711a8375f701c7c9e"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.620874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lfqcg" event={"ID":"a09631ae-f0ad-46af-b441-76ce38840f42","Type":"ContainerStarted","Data":"e224285167208d1794ca12e89898ba75c2d414c7235023cc2c422dce1790761a"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.623732 4726 generic.go:334] "Generic (PLEG): container finished" podID="e54464f6-5188-4631-a0d3-609ce5b2584c" containerID="b45e5b6955e72e5c0a7286fef2f30799d7b2f85c701556ce5a6e8e73ee78b41a" exitCode=0 Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.623808 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xszts" event={"ID":"e54464f6-5188-4631-a0d3-609ce5b2584c","Type":"ContainerDied","Data":"b45e5b6955e72e5c0a7286fef2f30799d7b2f85c701556ce5a6e8e73ee78b41a"} Oct 04 03:59:18 crc kubenswrapper[4726]: I1004 03:59:18.623846 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xszts" event={"ID":"e54464f6-5188-4631-a0d3-609ce5b2584c","Type":"ContainerStarted","Data":"e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b"} Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.010912 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148200 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148303 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148516 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zrkl\" (UniqueName: \"kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148542 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148625 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148656 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.148698 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn\") pod \"bd02c4f5-067a-47c7-9ab5-47316427397b\" (UID: \"bd02c4f5-067a-47c7-9ab5-47316427397b\") " Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.149015 4726 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.149051 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.149080 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run" (OuterVolumeSpecName: "var-run") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.149320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.150367 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts" (OuterVolumeSpecName: "scripts") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.163188 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5kwdw-config-xngdk"] Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.172263 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl" (OuterVolumeSpecName: "kube-api-access-2zrkl") pod "bd02c4f5-067a-47c7-9ab5-47316427397b" (UID: "bd02c4f5-067a-47c7-9ab5-47316427397b"). InnerVolumeSpecName "kube-api-access-2zrkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.173830 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5kwdw-config-xngdk"] Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.250575 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zrkl\" (UniqueName: \"kubernetes.io/projected/bd02c4f5-067a-47c7-9ab5-47316427397b-kube-api-access-2zrkl\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.250608 4726 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.250618 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd02c4f5-067a-47c7-9ab5-47316427397b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.250627 4726 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.250635 4726 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd02c4f5-067a-47c7-9ab5-47316427397b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.636442 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"158b6bed85086faca8e18f72edce55b190a956d0b18446250f95104df250b2cb"} Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.638139 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="518bb0483190be4530f8ef976f597ff0497ebcc85a0fa9bc11fdf5e3c172b03c" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.638212 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5kwdw-config-xngdk" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.683874 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d7ee-account-create-7t464"] Oct 04 03:59:19 crc kubenswrapper[4726]: E1004 03:59:19.684306 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd02c4f5-067a-47c7-9ab5-47316427397b" containerName="ovn-config" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.684325 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd02c4f5-067a-47c7-9ab5-47316427397b" containerName="ovn-config" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.684508 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd02c4f5-067a-47c7-9ab5-47316427397b" containerName="ovn-config" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.685231 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.687274 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.691756 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d7ee-account-create-7t464"] Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.757396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdwqj\" (UniqueName: \"kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj\") pod \"keystone-d7ee-account-create-7t464\" (UID: \"d99bb218-074c-4ffd-abc8-a40bbbb2c611\") " pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.860719 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwqj\" (UniqueName: \"kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj\") pod \"keystone-d7ee-account-create-7t464\" (UID: \"d99bb218-074c-4ffd-abc8-a40bbbb2c611\") " pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.880254 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwqj\" (UniqueName: \"kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj\") pod \"keystone-d7ee-account-create-7t464\" (UID: \"d99bb218-074c-4ffd-abc8-a40bbbb2c611\") " pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.977096 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-987d-account-create-qs5z8"] Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.979184 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.981735 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 03:59:19 crc kubenswrapper[4726]: I1004 03:59:19.988883 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-987d-account-create-qs5z8"] Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.007968 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.064366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6m6d\" (UniqueName: \"kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d\") pod \"placement-987d-account-create-qs5z8\" (UID: \"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec\") " pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.172525 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6m6d\" (UniqueName: \"kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d\") pod \"placement-987d-account-create-qs5z8\" (UID: \"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec\") " pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.186115 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.199953 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6m6d\" (UniqueName: \"kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d\") pod \"placement-987d-account-create-qs5z8\" (UID: \"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec\") " pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.214633 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.220661 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xszts" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.273624 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c2xp\" (UniqueName: \"kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp\") pod \"e54464f6-5188-4631-a0d3-609ce5b2584c\" (UID: \"e54464f6-5188-4631-a0d3-609ce5b2584c\") " Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.273882 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4gh7\" (UniqueName: \"kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7\") pod \"cdee62dd-66dd-42da-840c-520048edfd3e\" (UID: \"cdee62dd-66dd-42da-840c-520048edfd3e\") " Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.273935 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfgm8\" (UniqueName: \"kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8\") pod \"a09631ae-f0ad-46af-b441-76ce38840f42\" (UID: \"a09631ae-f0ad-46af-b441-76ce38840f42\") " Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.277467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7" (OuterVolumeSpecName: "kube-api-access-c4gh7") pod "cdee62dd-66dd-42da-840c-520048edfd3e" (UID: "cdee62dd-66dd-42da-840c-520048edfd3e"). InnerVolumeSpecName "kube-api-access-c4gh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.278741 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8" (OuterVolumeSpecName: "kube-api-access-wfgm8") pod "a09631ae-f0ad-46af-b441-76ce38840f42" (UID: "a09631ae-f0ad-46af-b441-76ce38840f42"). InnerVolumeSpecName "kube-api-access-wfgm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.279985 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp" (OuterVolumeSpecName: "kube-api-access-7c2xp") pod "e54464f6-5188-4631-a0d3-609ce5b2584c" (UID: "e54464f6-5188-4631-a0d3-609ce5b2584c"). InnerVolumeSpecName "kube-api-access-7c2xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.305811 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.376194 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4gh7\" (UniqueName: \"kubernetes.io/projected/cdee62dd-66dd-42da-840c-520048edfd3e-kube-api-access-c4gh7\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.376228 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfgm8\" (UniqueName: \"kubernetes.io/projected/a09631ae-f0ad-46af-b441-76ce38840f42-kube-api-access-wfgm8\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.376239 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c2xp\" (UniqueName: \"kubernetes.io/projected/e54464f6-5188-4631-a0d3-609ce5b2584c-kube-api-access-7c2xp\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.539499 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd02c4f5-067a-47c7-9ab5-47316427397b" path="/var/lib/kubelet/pods/bd02c4f5-067a-47c7-9ab5-47316427397b/volumes" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.542095 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d7ee-account-create-7t464"] Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.646888 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d7ee-account-create-7t464" event={"ID":"d99bb218-074c-4ffd-abc8-a40bbbb2c611","Type":"ContainerStarted","Data":"1cb99923ceb886ff8a4dfc7008fe4e644598339f97766b0dc0c0f6c5942581ed"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.650934 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"01aa5b2edca1fbe80d8b9fc070012f7691212672a1911e7a18ee51d5e68f6891"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.650976 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"67feb6374035812734f3afc3f66bc1b38600682aa3585488329bb370f0daaf9e"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.658742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lfqcg" event={"ID":"a09631ae-f0ad-46af-b441-76ce38840f42","Type":"ContainerDied","Data":"e224285167208d1794ca12e89898ba75c2d414c7235023cc2c422dce1790761a"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.658761 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lfqcg" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.658778 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e224285167208d1794ca12e89898ba75c2d414c7235023cc2c422dce1790761a" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.662899 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xszts" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.663037 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xszts" event={"ID":"e54464f6-5188-4631-a0d3-609ce5b2584c","Type":"ContainerDied","Data":"e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.663078 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e47af38edf8bea3f2004eedd5cc94dcd516e91b8a818e2b7cfff0619be82d23b" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.668953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlls9" event={"ID":"cdee62dd-66dd-42da-840c-520048edfd3e","Type":"ContainerDied","Data":"967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208"} Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.668989 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="967205c9d2745e1260e2bf3ebd605e06a4a03d67a2bf83d493791afcbb9ee208" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.669010 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlls9" Oct 04 03:59:20 crc kubenswrapper[4726]: I1004 03:59:20.714403 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-987d-account-create-qs5z8"] Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.679206 4726 generic.go:334] "Generic (PLEG): container finished" podID="cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" containerID="7d38f7caecb6e1573027ffdabc8a8062865a86c0af63ba56756b2bba80978073" exitCode=0 Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.679416 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-987d-account-create-qs5z8" event={"ID":"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec","Type":"ContainerDied","Data":"7d38f7caecb6e1573027ffdabc8a8062865a86c0af63ba56756b2bba80978073"} Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.680787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-987d-account-create-qs5z8" event={"ID":"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec","Type":"ContainerStarted","Data":"838069c04cb55985baae4806ce2f29fa25df96bc77773eb4d22838e107fc9c73"} Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.684925 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"4bdd056c3cc4faeeef38d84a20321a8f8c973e53f9395b0d6feae0aaa84117e2"} Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.685151 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"e2b3671ae76ffc48c05c576d42ffc938ad98065ce6e530f482697d34611bd398"} Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.689004 4726 generic.go:334] "Generic (PLEG): container finished" podID="d99bb218-074c-4ffd-abc8-a40bbbb2c611" containerID="c7e1b8a524c53f78071e001f65e63c25b7dfb5133e67957143474537d36ac043" exitCode=0 Oct 04 03:59:21 crc kubenswrapper[4726]: I1004 03:59:21.689150 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d7ee-account-create-7t464" event={"ID":"d99bb218-074c-4ffd-abc8-a40bbbb2c611","Type":"ContainerDied","Data":"c7e1b8a524c53f78071e001f65e63c25b7dfb5133e67957143474537d36ac043"} Oct 04 03:59:22 crc kubenswrapper[4726]: I1004 03:59:22.702196 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"1cbf8e7b10b084fbf1e18a6460259d03675e87b26ce32aa6131efd7a1d2a77bd"} Oct 04 03:59:22 crc kubenswrapper[4726]: I1004 03:59:22.702230 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"4f6cd9219cac5571fb880b99397c5dd25c7df707e8904677040dacfec0a04c8e"} Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.973347 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8344-account-create-95zcb"] Oct 04 03:59:26 crc kubenswrapper[4726]: E1004 03:59:26.975648 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdee62dd-66dd-42da-840c-520048edfd3e" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.975673 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdee62dd-66dd-42da-840c-520048edfd3e" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: E1004 03:59:26.975700 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e54464f6-5188-4631-a0d3-609ce5b2584c" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.975712 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e54464f6-5188-4631-a0d3-609ce5b2584c" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: E1004 03:59:26.975757 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a09631ae-f0ad-46af-b441-76ce38840f42" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.975768 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a09631ae-f0ad-46af-b441-76ce38840f42" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.976006 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e54464f6-5188-4631-a0d3-609ce5b2584c" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.976033 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a09631ae-f0ad-46af-b441-76ce38840f42" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.976074 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdee62dd-66dd-42da-840c-520048edfd3e" containerName="mariadb-database-create" Oct 04 03:59:26 crc kubenswrapper[4726]: I1004 03:59:26.982991 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.001455 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8344-account-create-95zcb"] Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.005359 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.021182 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g8lv\" (UniqueName: \"kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv\") pod \"barbican-8344-account-create-95zcb\" (UID: \"ae06837e-2532-440c-9a0f-65dd2b70b1e7\") " pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.074784 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6ea2-account-create-8tn2m"] Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.076004 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.078259 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.083236 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6ea2-account-create-8tn2m"] Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.122426 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grrqm\" (UniqueName: \"kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm\") pod \"cinder-6ea2-account-create-8tn2m\" (UID: \"5383a9a4-9771-40c8-b9fb-4fcbe87deb85\") " pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.122532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g8lv\" (UniqueName: \"kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv\") pod \"barbican-8344-account-create-95zcb\" (UID: \"ae06837e-2532-440c-9a0f-65dd2b70b1e7\") " pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.142169 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g8lv\" (UniqueName: \"kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv\") pod \"barbican-8344-account-create-95zcb\" (UID: \"ae06837e-2532-440c-9a0f-65dd2b70b1e7\") " pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.173938 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-613a-account-create-ccqbc"] Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.175004 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.178040 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.181522 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-613a-account-create-ccqbc"] Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.224329 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s44b6\" (UniqueName: \"kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6\") pod \"neutron-613a-account-create-ccqbc\" (UID: \"19f850a7-b8df-4af9-a0df-0f78a6f37482\") " pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.224745 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grrqm\" (UniqueName: \"kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm\") pod \"cinder-6ea2-account-create-8tn2m\" (UID: \"5383a9a4-9771-40c8-b9fb-4fcbe87deb85\") " pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.241037 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grrqm\" (UniqueName: \"kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm\") pod \"cinder-6ea2-account-create-8tn2m\" (UID: \"5383a9a4-9771-40c8-b9fb-4fcbe87deb85\") " pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.303996 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.326422 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s44b6\" (UniqueName: \"kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6\") pod \"neutron-613a-account-create-ccqbc\" (UID: \"19f850a7-b8df-4af9-a0df-0f78a6f37482\") " pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.342029 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s44b6\" (UniqueName: \"kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6\") pod \"neutron-613a-account-create-ccqbc\" (UID: \"19f850a7-b8df-4af9-a0df-0f78a6f37482\") " pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.395844 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:27 crc kubenswrapper[4726]: I1004 03:59:27.505551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:29 crc kubenswrapper[4726]: I1004 03:59:29.981497 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.777711 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d7ee-account-create-7t464" event={"ID":"d99bb218-074c-4ffd-abc8-a40bbbb2c611","Type":"ContainerDied","Data":"1cb99923ceb886ff8a4dfc7008fe4e644598339f97766b0dc0c0f6c5942581ed"} Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.778073 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cb99923ceb886ff8a4dfc7008fe4e644598339f97766b0dc0c0f6c5942581ed" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.784366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-987d-account-create-qs5z8" event={"ID":"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec","Type":"ContainerDied","Data":"838069c04cb55985baae4806ce2f29fa25df96bc77773eb4d22838e107fc9c73"} Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.784421 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="838069c04cb55985baae4806ce2f29fa25df96bc77773eb4d22838e107fc9c73" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.850774 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.857214 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.899053 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6m6d\" (UniqueName: \"kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d\") pod \"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec\" (UID: \"cb49d8bc-c2a0-41ac-9a64-de6e6f502cec\") " Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.899602 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdwqj\" (UniqueName: \"kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj\") pod \"d99bb218-074c-4ffd-abc8-a40bbbb2c611\" (UID: \"d99bb218-074c-4ffd-abc8-a40bbbb2c611\") " Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.907552 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj" (OuterVolumeSpecName: "kube-api-access-rdwqj") pod "d99bb218-074c-4ffd-abc8-a40bbbb2c611" (UID: "d99bb218-074c-4ffd-abc8-a40bbbb2c611"). InnerVolumeSpecName "kube-api-access-rdwqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:30 crc kubenswrapper[4726]: I1004 03:59:30.908323 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d" (OuterVolumeSpecName: "kube-api-access-m6m6d") pod "cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" (UID: "cb49d8bc-c2a0-41ac-9a64-de6e6f502cec"). InnerVolumeSpecName "kube-api-access-m6m6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.001262 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6m6d\" (UniqueName: \"kubernetes.io/projected/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec-kube-api-access-m6m6d\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.001293 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdwqj\" (UniqueName: \"kubernetes.io/projected/d99bb218-074c-4ffd-abc8-a40bbbb2c611-kube-api-access-rdwqj\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.273837 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6ea2-account-create-8tn2m"] Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.287936 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-613a-account-create-ccqbc"] Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.293228 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8344-account-create-95zcb"] Oct 04 03:59:31 crc kubenswrapper[4726]: W1004 03:59:31.295819 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19f850a7_b8df_4af9_a0df_0f78a6f37482.slice/crio-23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2 WatchSource:0}: Error finding container 23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2: Status 404 returned error can't find the container with id 23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2 Oct 04 03:59:31 crc kubenswrapper[4726]: W1004 03:59:31.299794 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae06837e_2532_440c_9a0f_65dd2b70b1e7.slice/crio-e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84 WatchSource:0}: Error finding container e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84: Status 404 returned error can't find the container with id e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84 Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.796810 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"b0afd8d06937f9f4397f5cf44848ab061b7c2a84a3723f08d237da2e7adae1df"} Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.797605 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ea2-account-create-8tn2m" event={"ID":"5383a9a4-9771-40c8-b9fb-4fcbe87deb85","Type":"ContainerStarted","Data":"4621714999a13bf2399b7c7ed9f4bf1584d3d5a4b20bd6f0674c8866b52dedf5"} Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.798796 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8344-account-create-95zcb" event={"ID":"ae06837e-2532-440c-9a0f-65dd2b70b1e7","Type":"ContainerStarted","Data":"e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84"} Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.799751 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d7ee-account-create-7t464" Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.802218 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-987d-account-create-qs5z8" Oct 04 03:59:31 crc kubenswrapper[4726]: I1004 03:59:31.802208 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-613a-account-create-ccqbc" event={"ID":"19f850a7-b8df-4af9-a0df-0f78a6f37482","Type":"ContainerStarted","Data":"23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.823548 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4kwmz" event={"ID":"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a","Type":"ContainerStarted","Data":"5109ad53f6c94ca393fccc0dc0ee46d2f18a40f93b8cc828a299bef1970733c5"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.827515 4726 generic.go:334] "Generic (PLEG): container finished" podID="19f850a7-b8df-4af9-a0df-0f78a6f37482" containerID="c97e6108ede611f87f19cbf31291f274b83a33880ce0078930832fa81e5600b3" exitCode=0 Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.827614 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-613a-account-create-ccqbc" event={"ID":"19f850a7-b8df-4af9-a0df-0f78a6f37482","Type":"ContainerDied","Data":"c97e6108ede611f87f19cbf31291f274b83a33880ce0078930832fa81e5600b3"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.841287 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"7167b84f1616f16a9c885545ce727e9665fe29069f8ad1cf63d94ddabd9ba1fe"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.841341 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"1f41926f4ee43e7ad6bea69e01e99d46739f6c8274e26afd0fe2c6611021ba95"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.841392 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"1d57efae817fe32a5abd6ef1b0ef98c3547acc38ad357ce251a25efd9676d79a"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.849608 4726 generic.go:334] "Generic (PLEG): container finished" podID="5383a9a4-9771-40c8-b9fb-4fcbe87deb85" containerID="4420bcf7200cb7fd4a3cfea501e58e3e9b082cb9b59788d7278650033aa7230e" exitCode=0 Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.849930 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ea2-account-create-8tn2m" event={"ID":"5383a9a4-9771-40c8-b9fb-4fcbe87deb85","Type":"ContainerDied","Data":"4420bcf7200cb7fd4a3cfea501e58e3e9b082cb9b59788d7278650033aa7230e"} Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.851037 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4kwmz" podStartSLOduration=2.828668034 podStartE2EDuration="17.851016968s" podCreationTimestamp="2025-10-04 03:59:15 +0000 UTC" firstStartedPulling="2025-10-04 03:59:16.345642066 +0000 UTC m=+1130.520265279" lastFinishedPulling="2025-10-04 03:59:31.367991 +0000 UTC m=+1145.542614213" observedRunningTime="2025-10-04 03:59:32.850578808 +0000 UTC m=+1147.025202091" watchObservedRunningTime="2025-10-04 03:59:32.851016968 +0000 UTC m=+1147.025640211" Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.855458 4726 generic.go:334] "Generic (PLEG): container finished" podID="ae06837e-2532-440c-9a0f-65dd2b70b1e7" containerID="1b41125dcf3e5bb13869643424be34e7281b96914fe4b3dc5d14963d53ffce46" exitCode=0 Oct 04 03:59:32 crc kubenswrapper[4726]: I1004 03:59:32.855553 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8344-account-create-95zcb" event={"ID":"ae06837e-2532-440c-9a0f-65dd2b70b1e7","Type":"ContainerDied","Data":"1b41125dcf3e5bb13869643424be34e7281b96914fe4b3dc5d14963d53ffce46"} Oct 04 03:59:33 crc kubenswrapper[4726]: I1004 03:59:33.877259 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c4528cfc-3f8b-4850-aeaa-eabe209382bd","Type":"ContainerStarted","Data":"0137348c4acb8c96ac0ac2b5f4e66ead5fad2aa6e5573f149153beb5676e1f2c"} Oct 04 03:59:33 crc kubenswrapper[4726]: I1004 03:59:33.944291 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=45.670437518 podStartE2EDuration="51.944271568s" podCreationTimestamp="2025-10-04 03:58:42 +0000 UTC" firstStartedPulling="2025-10-04 03:59:16.064822211 +0000 UTC m=+1130.239445444" lastFinishedPulling="2025-10-04 03:59:22.338656281 +0000 UTC m=+1136.513279494" observedRunningTime="2025-10-04 03:59:33.937977644 +0000 UTC m=+1148.112600867" watchObservedRunningTime="2025-10-04 03:59:33.944271568 +0000 UTC m=+1148.118894781" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.189676 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.189714 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.267361 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 03:59:34 crc kubenswrapper[4726]: E1004 03:59:34.267719 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.267733 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: E1004 03:59:34.267767 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99bb218-074c-4ffd-abc8-a40bbbb2c611" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.267775 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99bb218-074c-4ffd-abc8-a40bbbb2c611" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.267944 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d99bb218-074c-4ffd-abc8-a40bbbb2c611" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.267956 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" containerName="mariadb-account-create" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.271875 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.275474 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.275728 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.278243 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368215 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s44b6\" (UniqueName: \"kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6\") pod \"19f850a7-b8df-4af9-a0df-0f78a6f37482\" (UID: \"19f850a7-b8df-4af9-a0df-0f78a6f37482\") " Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368529 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368610 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368630 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bvsz\" (UniqueName: \"kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.368664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.373919 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6" (OuterVolumeSpecName: "kube-api-access-s44b6") pod "19f850a7-b8df-4af9-a0df-0f78a6f37482" (UID: "19f850a7-b8df-4af9-a0df-0f78a6f37482"). InnerVolumeSpecName "kube-api-access-s44b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.409507 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.415890 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470142 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g8lv\" (UniqueName: \"kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv\") pod \"ae06837e-2532-440c-9a0f-65dd2b70b1e7\" (UID: \"ae06837e-2532-440c-9a0f-65dd2b70b1e7\") " Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470238 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grrqm\" (UniqueName: \"kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm\") pod \"5383a9a4-9771-40c8-b9fb-4fcbe87deb85\" (UID: \"5383a9a4-9771-40c8-b9fb-4fcbe87deb85\") " Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bvsz\" (UniqueName: \"kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470532 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470597 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470638 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470665 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470692 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.470772 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s44b6\" (UniqueName: \"kubernetes.io/projected/19f850a7-b8df-4af9-a0df-0f78a6f37482-kube-api-access-s44b6\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.471722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.471943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.472841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.473811 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv" (OuterVolumeSpecName: "kube-api-access-8g8lv") pod "ae06837e-2532-440c-9a0f-65dd2b70b1e7" (UID: "ae06837e-2532-440c-9a0f-65dd2b70b1e7"). InnerVolumeSpecName "kube-api-access-8g8lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.473904 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.473948 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.475632 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm" (OuterVolumeSpecName: "kube-api-access-grrqm") pod "5383a9a4-9771-40c8-b9fb-4fcbe87deb85" (UID: "5383a9a4-9771-40c8-b9fb-4fcbe87deb85"). InnerVolumeSpecName "kube-api-access-grrqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.488370 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bvsz\" (UniqueName: \"kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz\") pod \"dnsmasq-dns-6d5b6d6b67-4pcp4\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.572800 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g8lv\" (UniqueName: \"kubernetes.io/projected/ae06837e-2532-440c-9a0f-65dd2b70b1e7-kube-api-access-8g8lv\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.572834 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grrqm\" (UniqueName: \"kubernetes.io/projected/5383a9a4-9771-40c8-b9fb-4fcbe87deb85-kube-api-access-grrqm\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.588949 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.839303 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 03:59:34 crc kubenswrapper[4726]: W1004 03:59:34.843044 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec74c29b_8285_4194_8d19_87475b516be3.slice/crio-c1aed412aa452a0d6bc3d97b9610c42a529e273aa9e8f737580793220ed4d138 WatchSource:0}: Error finding container c1aed412aa452a0d6bc3d97b9610c42a529e273aa9e8f737580793220ed4d138: Status 404 returned error can't find the container with id c1aed412aa452a0d6bc3d97b9610c42a529e273aa9e8f737580793220ed4d138 Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.891224 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-613a-account-create-ccqbc" event={"ID":"19f850a7-b8df-4af9-a0df-0f78a6f37482","Type":"ContainerDied","Data":"23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2"} Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.891270 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23216836a0af7df74e87c4f5624d63ab1b69ec59a9b702726f307ff004cd38a2" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.891340 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-613a-account-create-ccqbc" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.907926 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" event={"ID":"ec74c29b-8285-4194-8d19-87475b516be3","Type":"ContainerStarted","Data":"c1aed412aa452a0d6bc3d97b9610c42a529e273aa9e8f737580793220ed4d138"} Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.918019 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6ea2-account-create-8tn2m" event={"ID":"5383a9a4-9771-40c8-b9fb-4fcbe87deb85","Type":"ContainerDied","Data":"4621714999a13bf2399b7c7ed9f4bf1584d3d5a4b20bd6f0674c8866b52dedf5"} Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.918370 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4621714999a13bf2399b7c7ed9f4bf1584d3d5a4b20bd6f0674c8866b52dedf5" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.918501 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6ea2-account-create-8tn2m" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.927452 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8344-account-create-95zcb" Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.928179 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8344-account-create-95zcb" event={"ID":"ae06837e-2532-440c-9a0f-65dd2b70b1e7","Type":"ContainerDied","Data":"e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84"} Oct 04 03:59:34 crc kubenswrapper[4726]: I1004 03:59:34.928236 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e42af6f2f66909560bc1a556f5c16f1a9a34a86c84ae39949c6b688550aaee84" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246283 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-p7f5d"] Oct 04 03:59:35 crc kubenswrapper[4726]: E1004 03:59:35.246613 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae06837e-2532-440c-9a0f-65dd2b70b1e7" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246629 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae06837e-2532-440c-9a0f-65dd2b70b1e7" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: E1004 03:59:35.246651 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f850a7-b8df-4af9-a0df-0f78a6f37482" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246658 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f850a7-b8df-4af9-a0df-0f78a6f37482" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: E1004 03:59:35.246667 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5383a9a4-9771-40c8-b9fb-4fcbe87deb85" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246673 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5383a9a4-9771-40c8-b9fb-4fcbe87deb85" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246826 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae06837e-2532-440c-9a0f-65dd2b70b1e7" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246844 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5383a9a4-9771-40c8-b9fb-4fcbe87deb85" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.246857 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f850a7-b8df-4af9-a0df-0f78a6f37482" containerName="mariadb-account-create" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.247327 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.249968 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.250177 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.250216 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.250601 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2nwhj" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.288311 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.288449 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.288506 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvx9t\" (UniqueName: \"kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.289278 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-p7f5d"] Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.390406 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvx9t\" (UniqueName: \"kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.390529 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.390609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.396996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.398892 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.409876 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvx9t\" (UniqueName: \"kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t\") pod \"keystone-db-sync-p7f5d\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.592541 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.935667 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec74c29b-8285-4194-8d19-87475b516be3" containerID="423b36e48db8be097e104487f0292e74540da76ede97b8b3cf9b07065e1e3b30" exitCode=0 Oct 04 03:59:35 crc kubenswrapper[4726]: I1004 03:59:35.935922 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" event={"ID":"ec74c29b-8285-4194-8d19-87475b516be3","Type":"ContainerDied","Data":"423b36e48db8be097e104487f0292e74540da76ede97b8b3cf9b07065e1e3b30"} Oct 04 03:59:36 crc kubenswrapper[4726]: I1004 03:59:36.054136 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-p7f5d"] Oct 04 03:59:36 crc kubenswrapper[4726]: I1004 03:59:36.944983 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p7f5d" event={"ID":"480a81a8-69b1-4894-b36c-9f65a681af26","Type":"ContainerStarted","Data":"0fb06b41ea16330a3919778a78db8bb5928f29dda99c81c0bf43d53d760a32ce"} Oct 04 03:59:36 crc kubenswrapper[4726]: I1004 03:59:36.946916 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" event={"ID":"ec74c29b-8285-4194-8d19-87475b516be3","Type":"ContainerStarted","Data":"5c03e6e615995fc8526418938595ff2e3646f2be5b0d46459c0e18672d1acb2d"} Oct 04 03:59:36 crc kubenswrapper[4726]: I1004 03:59:36.947155 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:36 crc kubenswrapper[4726]: I1004 03:59:36.973734 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podStartSLOduration=2.973716293 podStartE2EDuration="2.973716293s" podCreationTimestamp="2025-10-04 03:59:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:36.97096683 +0000 UTC m=+1151.145590043" watchObservedRunningTime="2025-10-04 03:59:36.973716293 +0000 UTC m=+1151.148339506" Oct 04 03:59:44 crc kubenswrapper[4726]: I1004 03:59:44.591033 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 03:59:44 crc kubenswrapper[4726]: I1004 03:59:44.656022 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:59:44 crc kubenswrapper[4726]: I1004 03:59:44.656316 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="dnsmasq-dns" containerID="cri-o://d8c5ab6e778403ee100c722bdc069e9a069661dfc083b54ecd9ffc4f7e0dc861" gracePeriod=10 Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.011048 4726 generic.go:334] "Generic (PLEG): container finished" podID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerID="d8c5ab6e778403ee100c722bdc069e9a069661dfc083b54ecd9ffc4f7e0dc861" exitCode=0 Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.011163 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" event={"ID":"e415f14e-2ec6-4e9c-89dc-df30dc0d7152","Type":"ContainerDied","Data":"d8c5ab6e778403ee100c722bdc069e9a069661dfc083b54ecd9ffc4f7e0dc861"} Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.259349 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.265755 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config\") pod \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.265842 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc\") pod \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.265892 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb\") pod \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.265907 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb\") pod \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.265930 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgdm4\" (UniqueName: \"kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4\") pod \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\" (UID: \"e415f14e-2ec6-4e9c-89dc-df30dc0d7152\") " Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.271323 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4" (OuterVolumeSpecName: "kube-api-access-hgdm4") pod "e415f14e-2ec6-4e9c-89dc-df30dc0d7152" (UID: "e415f14e-2ec6-4e9c-89dc-df30dc0d7152"). InnerVolumeSpecName "kube-api-access-hgdm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.323668 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e415f14e-2ec6-4e9c-89dc-df30dc0d7152" (UID: "e415f14e-2ec6-4e9c-89dc-df30dc0d7152"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.323681 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e415f14e-2ec6-4e9c-89dc-df30dc0d7152" (UID: "e415f14e-2ec6-4e9c-89dc-df30dc0d7152"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.325064 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e415f14e-2ec6-4e9c-89dc-df30dc0d7152" (UID: "e415f14e-2ec6-4e9c-89dc-df30dc0d7152"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.331847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config" (OuterVolumeSpecName: "config") pod "e415f14e-2ec6-4e9c-89dc-df30dc0d7152" (UID: "e415f14e-2ec6-4e9c-89dc-df30dc0d7152"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.367717 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.367750 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.367763 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.367776 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:45 crc kubenswrapper[4726]: I1004 03:59:45.367790 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgdm4\" (UniqueName: \"kubernetes.io/projected/e415f14e-2ec6-4e9c-89dc-df30dc0d7152-kube-api-access-hgdm4\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.022783 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.022797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-qjv7m" event={"ID":"e415f14e-2ec6-4e9c-89dc-df30dc0d7152","Type":"ContainerDied","Data":"4bda3405a87aad1783fd2ba123dcc5b8294d2bb947361391874c476605c32318"} Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.023297 4726 scope.go:117] "RemoveContainer" containerID="d8c5ab6e778403ee100c722bdc069e9a069661dfc083b54ecd9ffc4f7e0dc861" Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.027423 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p7f5d" event={"ID":"480a81a8-69b1-4894-b36c-9f65a681af26","Type":"ContainerStarted","Data":"06db8e023388e4bbb5cfc1bbf8ef2da4d68418cac61a93f0b40b5e26b32b7015"} Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.049370 4726 scope.go:117] "RemoveContainer" containerID="53f0171f7f26ba1a820ef9b5fcdee5dde1bcc0b09388bfe2e43bca8df5698a2e" Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.053682 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-p7f5d" podStartSLOduration=1.714851033 podStartE2EDuration="11.053642075s" podCreationTimestamp="2025-10-04 03:59:35 +0000 UTC" firstStartedPulling="2025-10-04 03:59:36.061349612 +0000 UTC m=+1150.235972845" lastFinishedPulling="2025-10-04 03:59:45.400140674 +0000 UTC m=+1159.574763887" observedRunningTime="2025-10-04 03:59:46.045907448 +0000 UTC m=+1160.220530671" watchObservedRunningTime="2025-10-04 03:59:46.053642075 +0000 UTC m=+1160.228265288" Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.070174 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.074471 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-qjv7m"] Oct 04 03:59:46 crc kubenswrapper[4726]: I1004 03:59:46.539316 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" path="/var/lib/kubelet/pods/e415f14e-2ec6-4e9c-89dc-df30dc0d7152/volumes" Oct 04 03:59:47 crc kubenswrapper[4726]: I1004 03:59:47.039183 4726 generic.go:334] "Generic (PLEG): container finished" podID="fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" containerID="5109ad53f6c94ca393fccc0dc0ee46d2f18a40f93b8cc828a299bef1970733c5" exitCode=0 Oct 04 03:59:47 crc kubenswrapper[4726]: I1004 03:59:47.039268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4kwmz" event={"ID":"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a","Type":"ContainerDied","Data":"5109ad53f6c94ca393fccc0dc0ee46d2f18a40f93b8cc828a299bef1970733c5"} Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.539878 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.624521 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data\") pod \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.624624 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data\") pod \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.624683 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle\") pod \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.624780 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jc4s\" (UniqueName: \"kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s\") pod \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\" (UID: \"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a\") " Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.631400 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" (UID: "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.631864 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s" (OuterVolumeSpecName: "kube-api-access-4jc4s") pod "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" (UID: "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a"). InnerVolumeSpecName "kube-api-access-4jc4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.658735 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" (UID: "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.705840 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data" (OuterVolumeSpecName: "config-data") pod "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" (UID: "fa0c7513-86d8-4b89-82ab-0a07c2d0d58a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.726667 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.726705 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.726716 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:48 crc kubenswrapper[4726]: I1004 03:59:48.726727 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jc4s\" (UniqueName: \"kubernetes.io/projected/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a-kube-api-access-4jc4s\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.064898 4726 generic.go:334] "Generic (PLEG): container finished" podID="480a81a8-69b1-4894-b36c-9f65a681af26" containerID="06db8e023388e4bbb5cfc1bbf8ef2da4d68418cac61a93f0b40b5e26b32b7015" exitCode=0 Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.065008 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p7f5d" event={"ID":"480a81a8-69b1-4894-b36c-9f65a681af26","Type":"ContainerDied","Data":"06db8e023388e4bbb5cfc1bbf8ef2da4d68418cac61a93f0b40b5e26b32b7015"} Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.068658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4kwmz" event={"ID":"fa0c7513-86d8-4b89-82ab-0a07c2d0d58a","Type":"ContainerDied","Data":"c822889ec6a2c293f0cd8cb77a37d926ef9116377854f91efbcfe0b99aa160d6"} Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.068701 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c822889ec6a2c293f0cd8cb77a37d926ef9116377854f91efbcfe0b99aa160d6" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.068756 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4kwmz" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.521003 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:49 crc kubenswrapper[4726]: E1004 03:59:49.521707 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" containerName="glance-db-sync" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.521728 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" containerName="glance-db-sync" Oct 04 03:59:49 crc kubenswrapper[4726]: E1004 03:59:49.521770 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="init" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.521778 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="init" Oct 04 03:59:49 crc kubenswrapper[4726]: E1004 03:59:49.521795 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="dnsmasq-dns" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.521803 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="dnsmasq-dns" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.521983 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" containerName="glance-db-sync" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.522015 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e415f14e-2ec6-4e9c-89dc-df30dc0d7152" containerName="dnsmasq-dns" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.523073 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.534243 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8mwr\" (UniqueName: \"kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548207 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548237 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548256 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.548275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649805 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649851 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649874 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649894 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8mwr\" (UniqueName: \"kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.649986 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.650720 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.650979 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.650982 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.651060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.651593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.673346 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8mwr\" (UniqueName: \"kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr\") pod \"dnsmasq-dns-895cf5cf-clr99\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:49 crc kubenswrapper[4726]: I1004 03:59:49.837558 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.281717 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:50 crc kubenswrapper[4726]: W1004 03:59:50.290272 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1109b03_df1b_4c27_8370_3e0d7028916d.slice/crio-21a52c404928d961fe5c6fc1a6b869c672e9d4904da4f0361dde3f7610f67e56 WatchSource:0}: Error finding container 21a52c404928d961fe5c6fc1a6b869c672e9d4904da4f0361dde3f7610f67e56: Status 404 returned error can't find the container with id 21a52c404928d961fe5c6fc1a6b869c672e9d4904da4f0361dde3f7610f67e56 Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.387440 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.462324 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle\") pod \"480a81a8-69b1-4894-b36c-9f65a681af26\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.462383 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data\") pod \"480a81a8-69b1-4894-b36c-9f65a681af26\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.462438 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvx9t\" (UniqueName: \"kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t\") pod \"480a81a8-69b1-4894-b36c-9f65a681af26\" (UID: \"480a81a8-69b1-4894-b36c-9f65a681af26\") " Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.466442 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t" (OuterVolumeSpecName: "kube-api-access-zvx9t") pod "480a81a8-69b1-4894-b36c-9f65a681af26" (UID: "480a81a8-69b1-4894-b36c-9f65a681af26"). InnerVolumeSpecName "kube-api-access-zvx9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.491601 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "480a81a8-69b1-4894-b36c-9f65a681af26" (UID: "480a81a8-69b1-4894-b36c-9f65a681af26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.512269 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data" (OuterVolumeSpecName: "config-data") pod "480a81a8-69b1-4894-b36c-9f65a681af26" (UID: "480a81a8-69b1-4894-b36c-9f65a681af26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.563864 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.563887 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480a81a8-69b1-4894-b36c-9f65a681af26-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:50 crc kubenswrapper[4726]: I1004 03:59:50.563896 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvx9t\" (UniqueName: \"kubernetes.io/projected/480a81a8-69b1-4894-b36c-9f65a681af26-kube-api-access-zvx9t\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.097468 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-p7f5d" event={"ID":"480a81a8-69b1-4894-b36c-9f65a681af26","Type":"ContainerDied","Data":"0fb06b41ea16330a3919778a78db8bb5928f29dda99c81c0bf43d53d760a32ce"} Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.097781 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fb06b41ea16330a3919778a78db8bb5928f29dda99c81c0bf43d53d760a32ce" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.097842 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-p7f5d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.100423 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerID="8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f" exitCode=0 Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.100450 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-clr99" event={"ID":"f1109b03-df1b-4c27-8370-3e0d7028916d","Type":"ContainerDied","Data":"8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f"} Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.100469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-clr99" event={"ID":"f1109b03-df1b-4c27-8370-3e0d7028916d","Type":"ContainerStarted","Data":"21a52c404928d961fe5c6fc1a6b869c672e9d4904da4f0361dde3f7610f67e56"} Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.335527 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.391576 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:51 crc kubenswrapper[4726]: E1004 03:59:51.391912 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480a81a8-69b1-4894-b36c-9f65a681af26" containerName="keystone-db-sync" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.391923 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="480a81a8-69b1-4894-b36c-9f65a681af26" containerName="keystone-db-sync" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.392096 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="480a81a8-69b1-4894-b36c-9f65a681af26" containerName="keystone-db-sync" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.392921 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.409602 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ghnzv"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.410866 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.419205 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2nwhj" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.419465 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.419569 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.424611 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.427400 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.439940 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ghnzv"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486050 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn4zq\" (UniqueName: \"kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486502 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486527 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486620 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486644 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486734 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486776 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486799 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486830 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486867 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.486909 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9twn6\" (UniqueName: \"kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.582646 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.586554 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587786 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587808 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9twn6\" (UniqueName: \"kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587899 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn4zq\" (UniqueName: \"kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587925 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587942 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587968 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.587989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.588612 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.588699 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.588956 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.588960 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.589020 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gmpmn" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.589180 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.589668 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.595746 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.595891 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.604005 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.611079 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.612257 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.619273 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.631528 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-72jth"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.631647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn4zq\" (UniqueName: \"kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq\") pod \"dnsmasq-dns-6c9c9f998c-lrhd7\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.632700 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.633521 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.638298 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.638584 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.638880 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9twn6\" (UniqueName: \"kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6\") pod \"keystone-bootstrap-ghnzv\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.640756 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dkv48" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.652371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.679067 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-72jth"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.699855 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.699904 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.699927 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.699969 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgjg\" (UniqueName: \"kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.699988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700013 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700055 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvgjj\" (UniqueName: \"kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700074 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700091 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700123 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.700141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.717308 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.720114 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.723323 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.725440 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.725605 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.738333 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ghnzv" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.749122 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.793816 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-q6284"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.794856 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.798157 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.798347 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jgmht" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.798451 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802272 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss25f\" (UniqueName: \"kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802304 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802367 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802388 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802432 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802494 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgjg\" (UniqueName: \"kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802515 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802535 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802561 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802577 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802627 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802642 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvgjj\" (UniqueName: \"kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802677 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802693 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.802710 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.807272 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.807544 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.817008 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.821691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.832317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.834714 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.837877 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.838907 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.841348 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.865880 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgjg\" (UniqueName: \"kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg\") pod \"cinder-db-sync-72jth\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.865954 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-q6284"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.874648 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvgjj\" (UniqueName: \"kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj\") pod \"horizon-7464dd7565-96l6d\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.903946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4b47\" (UniqueName: \"kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904008 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss25f\" (UniqueName: \"kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904052 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904083 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904162 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904189 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904203 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904267 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.904295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.915007 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-j9jwl"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.949645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.950911 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.951198 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.956934 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.960424 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.962405 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.965874 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.974856 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7464dd7565-96l6d" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.975417 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.975478 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.976020 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mxjcb" Oct 04 03:59:51 crc kubenswrapper[4726]: I1004 03:59:51.984820 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-72jth" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.009130 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014456 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014509 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4b47\" (UniqueName: \"kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhtwc\" (UniqueName: \"kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014646 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.014718 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.047809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss25f\" (UniqueName: \"kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f\") pod \"ceilometer-0\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " pod="openstack/ceilometer-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.073507 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j9jwl"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.076772 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.083490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.091634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4b47\" (UniqueName: \"kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47\") pod \"neutron-db-sync-q6284\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.095409 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q6284" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.095559 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.102048 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.116165 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xxqgf"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117179 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117256 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117287 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117310 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mn4c\" (UniqueName: \"kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117333 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhtwc\" (UniqueName: \"kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117356 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117381 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117386 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.117739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.123798 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.123985 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.124144 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xs8d7" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.124149 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.124676 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xxqgf"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.124724 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-clr99" event={"ID":"f1109b03-df1b-4c27-8370-3e0d7028916d","Type":"ContainerStarted","Data":"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24"} Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.124820 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.126247 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-895cf5cf-clr99" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="dnsmasq-dns" containerID="cri-o://cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24" gracePeriod=10 Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.128329 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.132191 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.135122 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.140651 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.142832 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8sqhw" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.142961 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.166466 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.168387 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.181376 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhtwc\" (UniqueName: \"kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc\") pod \"barbican-db-sync-j9jwl\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.205146 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.221538 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222386 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222440 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222459 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222509 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222526 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222542 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9fl4\" (UniqueName: \"kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222565 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mn4c\" (UniqueName: \"kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222600 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjz57\" (UniqueName: \"kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222619 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222648 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222682 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222700 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222721 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxtlh\" (UniqueName: \"kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222749 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222782 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222815 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222850 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222871 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.222889 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.223497 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.224004 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.224266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.244054 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.244437 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.245557 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.245658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mn4c\" (UniqueName: \"kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c\") pod \"horizon-7dbbf66849-ds92b\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.252480 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.255565 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.256509 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.274303 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-895cf5cf-clr99" podStartSLOduration=3.274285972 podStartE2EDuration="3.274285972s" podCreationTimestamp="2025-10-04 03:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:52.202003164 +0000 UTC m=+1166.376626377" watchObservedRunningTime="2025-10-04 03:59:52.274285972 +0000 UTC m=+1166.448909185" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.326927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxtlh\" (UniqueName: \"kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327534 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327614 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327760 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.327792 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.329666 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.329877 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.329923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.330338 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.330414 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.330517 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331062 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331196 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331314 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331341 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331445 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.331811 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9fl4\" (UniqueName: \"kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.333837 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjz57\" (UniqueName: \"kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.333927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.334891 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.337600 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.339051 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.342327 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.351914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.353856 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.356759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9fl4\" (UniqueName: \"kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.362758 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjz57\" (UniqueName: \"kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57\") pod \"dnsmasq-dns-57c957c4ff-d72hr\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.370615 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxtlh\" (UniqueName: \"kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.394328 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.394476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.401210 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.402241 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data\") pod \"placement-db-sync-xxqgf\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.402405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.419898 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9jwl" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437012 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437125 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437150 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437165 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfgrc\" (UniqueName: \"kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437190 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437244 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.437259 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.478497 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542387 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542536 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542637 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542758 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.542798 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfgrc\" (UniqueName: \"kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.543195 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.544497 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.544992 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.553861 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.554206 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.562868 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.581450 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxqgf" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.600529 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.622893 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.658271 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.673044 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfgrc\" (UniqueName: \"kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc\") pod \"glance-default-external-api-0\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " pod="openstack/glance-default-external-api-0" Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.744763 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.796119 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ghnzv"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.799377 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.807402 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-72jth"] Oct 04 03:59:52 crc kubenswrapper[4726]: W1004 03:59:52.828715 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71dec834_afb0_4ace_98b5_3be330ae5bd5.slice/crio-26b166ddfd1608e2b858f55b473e617de5ef444a811c56694a7ab9e7dde290c3 WatchSource:0}: Error finding container 26b166ddfd1608e2b858f55b473e617de5ef444a811c56694a7ab9e7dde290c3: Status 404 returned error can't find the container with id 26b166ddfd1608e2b858f55b473e617de5ef444a811c56694a7ab9e7dde290c3 Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.846637 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-q6284"] Oct 04 03:59:52 crc kubenswrapper[4726]: I1004 03:59:52.938621 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.020339 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.020809 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:59:53 crc kubenswrapper[4726]: W1004 03:59:53.033228 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67531077_c83d_49e5_b077_87d3671159f1.slice/crio-153d89033e70faf5f09403c4376fb6737982e2e6d55a22c3feaf1ee6bc8a32ca WatchSource:0}: Error finding container 153d89033e70faf5f09403c4376fb6737982e2e6d55a22c3feaf1ee6bc8a32ca: Status 404 returned error can't find the container with id 153d89033e70faf5f09403c4376fb6737982e2e6d55a22c3feaf1ee6bc8a32ca Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.141152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q6284" event={"ID":"ba7d46cf-02bd-4674-93e9-b339da1b568d","Type":"ContainerStarted","Data":"b7c36f662d7e2525c2de8c24a7965ddcb8652c00c8c43b0eab2313f2c9249e39"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.142437 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerStarted","Data":"153d89033e70faf5f09403c4376fb6737982e2e6d55a22c3feaf1ee6bc8a32ca"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.143245 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ghnzv" event={"ID":"f320c18f-07fe-4255-a283-8128e0719f6a","Type":"ContainerStarted","Data":"77a0cc3237a8e5c71a92acb375c4fff41f941090b0ae68d75a9aa6c96c9713c6"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.144321 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerStarted","Data":"26b166ddfd1608e2b858f55b473e617de5ef444a811c56694a7ab9e7dde290c3"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.146587 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerID="cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24" exitCode=0 Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.146633 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-clr99" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.146635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-clr99" event={"ID":"f1109b03-df1b-4c27-8370-3e0d7028916d","Type":"ContainerDied","Data":"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.146729 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-clr99" event={"ID":"f1109b03-df1b-4c27-8370-3e0d7028916d","Type":"ContainerDied","Data":"21a52c404928d961fe5c6fc1a6b869c672e9d4904da4f0361dde3f7610f67e56"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.146744 4726 scope.go:117] "RemoveContainer" containerID="cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.151551 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-72jth" event={"ID":"1f5261e9-4ede-4c36-83cd-58cfe6176333","Type":"ContainerStarted","Data":"702795c02ae3ca02fcc6158cadaf5017e7822629eee9bae32a3d54a06c1ac5a7"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.153484 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" event={"ID":"5039be6d-2515-4d0b-9d10-83dc2f41a47e","Type":"ContainerStarted","Data":"dcf95d935c39a6df48e226626714c75735b975fe4f7b73caeb541781588f5126"} Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.158631 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8mwr\" (UniqueName: \"kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.158852 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.158992 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.159079 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.159231 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.159380 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0\") pod \"f1109b03-df1b-4c27-8370-3e0d7028916d\" (UID: \"f1109b03-df1b-4c27-8370-3e0d7028916d\") " Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.164709 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr" (OuterVolumeSpecName: "kube-api-access-f8mwr") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "kube-api-access-f8mwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.169891 4726 scope.go:117] "RemoveContainer" containerID="8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.206963 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j9jwl"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.234847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.247330 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.247983 4726 scope.go:117] "RemoveContainer" containerID="cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.248162 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: E1004 03:59:53.248303 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24\": container with ID starting with cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24 not found: ID does not exist" containerID="cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.248345 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24"} err="failed to get container status \"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24\": rpc error: code = NotFound desc = could not find container \"cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24\": container with ID starting with cd261f85c605147a7bb8374e2ad1b3b491337163207b2f0fe0d217f957df4c24 not found: ID does not exist" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.248374 4726 scope.go:117] "RemoveContainer" containerID="8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f" Oct 04 03:59:53 crc kubenswrapper[4726]: E1004 03:59:53.249290 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f\": container with ID starting with 8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f not found: ID does not exist" containerID="8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.249352 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f"} err="failed to get container status \"8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f\": rpc error: code = NotFound desc = could not find container \"8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f\": container with ID starting with 8654eb547f296f76cc22c1d0f7e89cf27e3fffb13e68bc96eacb35ac3b0f352f not found: ID does not exist" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.250099 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config" (OuterVolumeSpecName: "config") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.266867 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8mwr\" (UniqueName: \"kubernetes.io/projected/f1109b03-df1b-4c27-8370-3e0d7028916d-kube-api-access-f8mwr\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.266898 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.266907 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.266915 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.266923 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.269664 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f1109b03-df1b-4c27-8370-3e0d7028916d" (UID: "f1109b03-df1b-4c27-8370-3e0d7028916d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.368243 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1109b03-df1b-4c27-8370-3e0d7028916d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.383230 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.393927 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.401478 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xxqgf"] Oct 04 03:59:53 crc kubenswrapper[4726]: W1004 03:59:53.406276 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf9ee75_66f7_4371_9adc_c0ad0249698c.slice/crio-cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93 WatchSource:0}: Error finding container cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93: Status 404 returned error can't find the container with id cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93 Oct 04 03:59:53 crc kubenswrapper[4726]: W1004 03:59:53.409374 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bc61ad6_c85f_45e3_9167_095574996c61.slice/crio-8a15b5a1e87576a121dd5db87b85b018355c4e0f513a8af3dfa231ab4b61049f WatchSource:0}: Error finding container 8a15b5a1e87576a121dd5db87b85b018355c4e0f513a8af3dfa231ab4b61049f: Status 404 returned error can't find the container with id 8a15b5a1e87576a121dd5db87b85b018355c4e0f513a8af3dfa231ab4b61049f Oct 04 03:59:53 crc kubenswrapper[4726]: W1004 03:59:53.414582 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32f117f2_07b4_4f8b_9b51_63b1e7ffb97f.slice/crio-fc20dc8ee4a30c26dc877fdacdc6c5d7c635114885f3f643c580f666177ea268 WatchSource:0}: Error finding container fc20dc8ee4a30c26dc877fdacdc6c5d7c635114885f3f643c580f666177ea268: Status 404 returned error can't find the container with id fc20dc8ee4a30c26dc877fdacdc6c5d7c635114885f3f643c580f666177ea268 Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.484755 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.496561 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-clr99"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.627724 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:59:53 crc kubenswrapper[4726]: I1004 03:59:53.729461 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:59:53 crc kubenswrapper[4726]: W1004 03:59:53.798895 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e592e7_e612_4c3f_912a_b15e701f0a4d.slice/crio-8af473e97dc9d3a6b72426f71b41c72d0a40519d6036fb0c43d14af57e402602 WatchSource:0}: Error finding container 8af473e97dc9d3a6b72426f71b41c72d0a40519d6036fb0c43d14af57e402602: Status 404 returned error can't find the container with id 8af473e97dc9d3a6b72426f71b41c72d0a40519d6036fb0c43d14af57e402602 Oct 04 03:59:54 crc kubenswrapper[4726]: E1004 03:59:54.011378 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bc61ad6_c85f_45e3_9167_095574996c61.slice/crio-conmon-d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bc61ad6_c85f_45e3_9167_095574996c61.slice/crio-d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8.scope\": RecentStats: unable to find data in memory cache]" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.164319 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ghnzv" event={"ID":"f320c18f-07fe-4255-a283-8128e0719f6a","Type":"ContainerStarted","Data":"c788b936e0dbd761637811f8d0c7d0930ef0ed905f894ab01f13cc9839e622ca"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.178788 4726 generic.go:334] "Generic (PLEG): container finished" podID="2bc61ad6-c85f-45e3-9167-095574996c61" containerID="d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8" exitCode=0 Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.179968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" event={"ID":"2bc61ad6-c85f-45e3-9167-095574996c61","Type":"ContainerDied","Data":"d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.179990 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" event={"ID":"2bc61ad6-c85f-45e3-9167-095574996c61","Type":"ContainerStarted","Data":"8a15b5a1e87576a121dd5db87b85b018355c4e0f513a8af3dfa231ab4b61049f"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.184793 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ghnzv" podStartSLOduration=3.184774778 podStartE2EDuration="3.184774778s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:54.177463931 +0000 UTC m=+1168.352087144" watchObservedRunningTime="2025-10-04 03:59:54.184774778 +0000 UTC m=+1168.359397991" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.185941 4726 generic.go:334] "Generic (PLEG): container finished" podID="5039be6d-2515-4d0b-9d10-83dc2f41a47e" containerID="2fb785db41b89a6e53cd6b3a64ed58aaed22cda468b1e43a58c26fc526a6d953" exitCode=0 Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.186000 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" event={"ID":"5039be6d-2515-4d0b-9d10-83dc2f41a47e","Type":"ContainerDied","Data":"2fb785db41b89a6e53cd6b3a64ed58aaed22cda468b1e43a58c26fc526a6d953"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.189026 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q6284" event={"ID":"ba7d46cf-02bd-4674-93e9-b339da1b568d","Type":"ContainerStarted","Data":"36ed94a606fbe6d95569c338bfc829a027d7fbb9bdf4294663ce49b3b96b58b0"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.191055 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerStarted","Data":"2daea5937940635342fd2b19b15f3d2317b995fcd5702a73f63fcddefc577cfb"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.192315 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxqgf" event={"ID":"caf9ee75-66f7-4371-9adc-c0ad0249698c","Type":"ContainerStarted","Data":"cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.200968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9jwl" event={"ID":"90894cee-ffab-4485-a221-703de154182f","Type":"ContainerStarted","Data":"ec0b089df2ef265f090cf72d8935f68ece32401e2e9efec525ce7ff8f5d318ed"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.203879 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerStarted","Data":"8af473e97dc9d3a6b72426f71b41c72d0a40519d6036fb0c43d14af57e402602"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.214968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dbbf66849-ds92b" event={"ID":"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f","Type":"ContainerStarted","Data":"fc20dc8ee4a30c26dc877fdacdc6c5d7c635114885f3f643c580f666177ea268"} Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.240276 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-q6284" podStartSLOduration=3.240261154 podStartE2EDuration="3.240261154s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:54.230184472 +0000 UTC m=+1168.404807685" watchObservedRunningTime="2025-10-04 03:59:54.240261154 +0000 UTC m=+1168.414884367" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.528355 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" path="/var/lib/kubelet/pods/f1109b03-df1b-4c27-8370-3e0d7028916d/volumes" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.752820 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896280 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896423 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn4zq\" (UniqueName: \"kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896478 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896516 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896541 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.896603 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0\") pod \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\" (UID: \"5039be6d-2515-4d0b-9d10-83dc2f41a47e\") " Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.900516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq" (OuterVolumeSpecName: "kube-api-access-dn4zq") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "kube-api-access-dn4zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.921572 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.943590 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.958445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.971241 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config" (OuterVolumeSpecName: "config") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.973595 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5039be6d-2515-4d0b-9d10-83dc2f41a47e" (UID: "5039be6d-2515-4d0b-9d10-83dc2f41a47e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.998483 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-config\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.998952 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.998963 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.998972 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn4zq\" (UniqueName: \"kubernetes.io/projected/5039be6d-2515-4d0b-9d10-83dc2f41a47e-kube-api-access-dn4zq\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.998982 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:54 crc kubenswrapper[4726]: I1004 03:59:54.999006 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5039be6d-2515-4d0b-9d10-83dc2f41a47e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.252122 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.252194 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-lrhd7" event={"ID":"5039be6d-2515-4d0b-9d10-83dc2f41a47e","Type":"ContainerDied","Data":"dcf95d935c39a6df48e226626714c75735b975fe4f7b73caeb541781588f5126"} Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.252281 4726 scope.go:117] "RemoveContainer" containerID="2fb785db41b89a6e53cd6b3a64ed58aaed22cda468b1e43a58c26fc526a6d953" Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.256486 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerStarted","Data":"235a4cd8da81c16a981fd44492ab3327ffb2e8e5a2475de19dea3237633099e0"} Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.258853 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" event={"ID":"2bc61ad6-c85f-45e3-9167-095574996c61","Type":"ContainerStarted","Data":"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107"} Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.258901 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.262654 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerStarted","Data":"5374d5402ff685151858fe00a7ce6245f7cf3ce18874401b7e2db3ba1f9c2b82"} Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.282422 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" podStartSLOduration=4.282406589 podStartE2EDuration="4.282406589s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:55.282250255 +0000 UTC m=+1169.456873478" watchObservedRunningTime="2025-10-04 03:59:55.282406589 +0000 UTC m=+1169.457029802" Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.352928 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:55 crc kubenswrapper[4726]: I1004 03:59:55.355602 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-lrhd7"] Oct 04 03:59:56 crc kubenswrapper[4726]: I1004 03:59:56.277620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerStarted","Data":"8d6502612a8d96d2c27a59c733ba8d71d49002827af52e465691504bbfaadcce"} Oct 04 03:59:56 crc kubenswrapper[4726]: I1004 03:59:56.282254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerStarted","Data":"469e99b3af120fafd9dbe265bd5f91396f2a3c1ddac0a296db4f7f6729d5b633"} Oct 04 03:59:56 crc kubenswrapper[4726]: I1004 03:59:56.301201 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.301185275 podStartE2EDuration="5.301185275s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:56.29838898 +0000 UTC m=+1170.473012203" watchObservedRunningTime="2025-10-04 03:59:56.301185275 +0000 UTC m=+1170.475808488" Oct 04 03:59:56 crc kubenswrapper[4726]: I1004 03:59:56.525844 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5039be6d-2515-4d0b-9d10-83dc2f41a47e" path="/var/lib/kubelet/pods/5039be6d-2515-4d0b-9d10-83dc2f41a47e/volumes" Oct 04 03:59:56 crc kubenswrapper[4726]: I1004 03:59:56.526228 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.52620925 podStartE2EDuration="4.52620925s" podCreationTimestamp="2025-10-04 03:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 03:59:56.331192234 +0000 UTC m=+1170.505815447" watchObservedRunningTime="2025-10-04 03:59:56.52620925 +0000 UTC m=+1170.700832463" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.601725 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.690039 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.723487 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.736758 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.766245 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 03:59:57 crc kubenswrapper[4726]: E1004 03:59:57.766705 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="dnsmasq-dns" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.766728 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="dnsmasq-dns" Oct 04 03:59:57 crc kubenswrapper[4726]: E1004 03:59:57.766743 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="init" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.766751 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="init" Oct 04 03:59:57 crc kubenswrapper[4726]: E1004 03:59:57.766770 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5039be6d-2515-4d0b-9d10-83dc2f41a47e" containerName="init" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.766778 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5039be6d-2515-4d0b-9d10-83dc2f41a47e" containerName="init" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.766992 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1109b03-df1b-4c27-8370-3e0d7028916d" containerName="dnsmasq-dns" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.767024 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5039be6d-2515-4d0b-9d10-83dc2f41a47e" containerName="init" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.775367 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.782634 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.863313 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.863402 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.863470 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.863561 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.863690 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mns7z\" (UniqueName: \"kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.965622 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mns7z\" (UniqueName: \"kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.965734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.965771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.965811 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.966552 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.967375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.967491 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.968082 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.974597 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:57 crc kubenswrapper[4726]: I1004 03:59:57.983813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mns7z\" (UniqueName: \"kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z\") pod \"horizon-5cd8c86949-xgpbz\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.110367 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304357 4726 generic.go:334] "Generic (PLEG): container finished" podID="f320c18f-07fe-4255-a283-8128e0719f6a" containerID="c788b936e0dbd761637811f8d0c7d0930ef0ed905f894ab01f13cc9839e622ca" exitCode=0 Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304476 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ghnzv" event={"ID":"f320c18f-07fe-4255-a283-8128e0719f6a","Type":"ContainerDied","Data":"c788b936e0dbd761637811f8d0c7d0930ef0ed905f894ab01f13cc9839e622ca"} Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304550 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-log" containerID="cri-o://5374d5402ff685151858fe00a7ce6245f7cf3ce18874401b7e2db3ba1f9c2b82" gracePeriod=30 Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304640 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-httpd" containerID="cri-o://469e99b3af120fafd9dbe265bd5f91396f2a3c1ddac0a296db4f7f6729d5b633" gracePeriod=30 Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304918 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-log" containerID="cri-o://235a4cd8da81c16a981fd44492ab3327ffb2e8e5a2475de19dea3237633099e0" gracePeriod=30 Oct 04 03:59:58 crc kubenswrapper[4726]: I1004 03:59:58.304948 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-httpd" containerID="cri-o://8d6502612a8d96d2c27a59c733ba8d71d49002827af52e465691504bbfaadcce" gracePeriod=30 Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.316655 4726 generic.go:334] "Generic (PLEG): container finished" podID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerID="8d6502612a8d96d2c27a59c733ba8d71d49002827af52e465691504bbfaadcce" exitCode=0 Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.316972 4726 generic.go:334] "Generic (PLEG): container finished" podID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerID="235a4cd8da81c16a981fd44492ab3327ffb2e8e5a2475de19dea3237633099e0" exitCode=143 Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.316761 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerDied","Data":"8d6502612a8d96d2c27a59c733ba8d71d49002827af52e465691504bbfaadcce"} Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.317047 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerDied","Data":"235a4cd8da81c16a981fd44492ab3327ffb2e8e5a2475de19dea3237633099e0"} Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.321140 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerID="469e99b3af120fafd9dbe265bd5f91396f2a3c1ddac0a296db4f7f6729d5b633" exitCode=0 Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.321186 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerID="5374d5402ff685151858fe00a7ce6245f7cf3ce18874401b7e2db3ba1f9c2b82" exitCode=143 Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.321386 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerDied","Data":"469e99b3af120fafd9dbe265bd5f91396f2a3c1ddac0a296db4f7f6729d5b633"} Oct 04 03:59:59 crc kubenswrapper[4726]: I1004 03:59:59.321426 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerDied","Data":"5374d5402ff685151858fe00a7ce6245f7cf3ce18874401b7e2db3ba1f9c2b82"} Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.140411 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6"] Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.142215 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.144497 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.144736 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.151873 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6"] Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.212785 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.212831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.213095 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xrt\" (UniqueName: \"kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.314778 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.314827 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.314898 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xrt\" (UniqueName: \"kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.315911 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.325616 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.336388 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xrt\" (UniqueName: \"kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt\") pod \"collect-profiles-29325840-nc6h6\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:00 crc kubenswrapper[4726]: I1004 04:00:00.472095 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:02 crc kubenswrapper[4726]: I1004 04:00:02.626307 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 04:00:02 crc kubenswrapper[4726]: I1004 04:00:02.719136 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 04:00:02 crc kubenswrapper[4726]: I1004 04:00:02.719412 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" containerID="cri-o://5c03e6e615995fc8526418938595ff2e3646f2be5b0d46459c0e18672d1acb2d" gracePeriod=10 Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.359725 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.370941 4726 generic.go:334] "Generic (PLEG): container finished" podID="ec74c29b-8285-4194-8d19-87475b516be3" containerID="5c03e6e615995fc8526418938595ff2e3646f2be5b0d46459c0e18672d1acb2d" exitCode=0 Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.370993 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" event={"ID":"ec74c29b-8285-4194-8d19-87475b516be3","Type":"ContainerDied","Data":"5c03e6e615995fc8526418938595ff2e3646f2be5b0d46459c0e18672d1acb2d"} Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.403472 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.405229 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.408543 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.423016 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469434 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6dqb\" (UniqueName: \"kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469542 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469753 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469861 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469946 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.469974 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.473001 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.512956 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7786888bd-6khpf"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.515085 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.532620 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7786888bd-6khpf"] Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572053 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-secret-key\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572101 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/888736e3-9e3d-402b-9ea3-79a0ba740fe7-logs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572145 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-tls-certs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572178 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6dqb\" (UniqueName: \"kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572194 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572228 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wltg4\" (UniqueName: \"kubernetes.io/projected/888736e3-9e3d-402b-9ea3-79a0ba740fe7-kube-api-access-wltg4\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572264 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572285 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-config-data\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-scripts\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572321 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-combined-ca-bundle\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572349 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.572810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.574195 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.575202 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.580862 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.582519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.595658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.600165 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6dqb\" (UniqueName: \"kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb\") pod \"horizon-6b6cffb848-k9fxk\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678139 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-secret-key\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678466 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/888736e3-9e3d-402b-9ea3-79a0ba740fe7-logs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-tls-certs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678545 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wltg4\" (UniqueName: \"kubernetes.io/projected/888736e3-9e3d-402b-9ea3-79a0ba740fe7-kube-api-access-wltg4\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678586 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-config-data\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678605 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-scripts\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.678621 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-combined-ca-bundle\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.680581 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/888736e3-9e3d-402b-9ea3-79a0ba740fe7-logs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.681639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-config-data\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.682302 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/888736e3-9e3d-402b-9ea3-79a0ba740fe7-scripts\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.702405 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-tls-certs\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.707809 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-combined-ca-bundle\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.716932 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/888736e3-9e3d-402b-9ea3-79a0ba740fe7-horizon-secret-key\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.733765 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.741892 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wltg4\" (UniqueName: \"kubernetes.io/projected/888736e3-9e3d-402b-9ea3-79a0ba740fe7-kube-api-access-wltg4\") pod \"horizon-7786888bd-6khpf\" (UID: \"888736e3-9e3d-402b-9ea3-79a0ba740fe7\") " pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:03 crc kubenswrapper[4726]: I1004 04:00:03.834299 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.188613 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.188674 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.188722 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.189334 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.189388 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea" gracePeriod=600 Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.390481 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea" exitCode=0 Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.390545 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea"} Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.390591 4726 scope.go:117] "RemoveContainer" containerID="aaedf75e7975e29ea015c55b9f96ceae75bdbdaf1329420100d26a2185b27ee7" Oct 04 04:00:04 crc kubenswrapper[4726]: I1004 04:00:04.590904 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Oct 04 04:00:09 crc kubenswrapper[4726]: I1004 04:00:09.591649 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Oct 04 04:00:14 crc kubenswrapper[4726]: I1004 04:00:14.591624 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Oct 04 04:00:14 crc kubenswrapper[4726]: I1004 04:00:14.592436 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 04:00:22 crc kubenswrapper[4726]: I1004 04:00:22.601720 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:22 crc kubenswrapper[4726]: I1004 04:00:22.602559 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:22 crc kubenswrapper[4726]: I1004 04:00:22.939261 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:00:22 crc kubenswrapper[4726]: I1004 04:00:22.939330 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:00:23 crc kubenswrapper[4726]: E1004 04:00:23.775592 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 04 04:00:23 crc kubenswrapper[4726]: E1004 04:00:23.776488 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gxtlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xxqgf_openstack(caf9ee75-66f7-4371-9adc-c0ad0249698c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:00:23 crc kubenswrapper[4726]: E1004 04:00:23.778223 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xxqgf" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" Oct 04 04:00:24 crc kubenswrapper[4726]: E1004 04:00:24.576869 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xxqgf" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" Oct 04 04:00:24 crc kubenswrapper[4726]: I1004 04:00:24.592281 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.615174 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ghnzv" event={"ID":"f320c18f-07fe-4255-a283-8128e0719f6a","Type":"ContainerDied","Data":"77a0cc3237a8e5c71a92acb375c4fff41f941090b0ae68d75a9aa6c96c9713c6"} Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.615668 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77a0cc3237a8e5c71a92acb375c4fff41f941090b0ae68d75a9aa6c96c9713c6" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.628485 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ghnzv" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780313 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780488 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780544 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780633 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9twn6\" (UniqueName: \"kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.780683 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle\") pod \"f320c18f-07fe-4255-a283-8128e0719f6a\" (UID: \"f320c18f-07fe-4255-a283-8128e0719f6a\") " Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.786362 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.786440 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts" (OuterVolumeSpecName: "scripts") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.798268 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.871633 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.871762 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6" (OuterVolumeSpecName: "kube-api-access-9twn6") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "kube-api-access-9twn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.876268 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data" (OuterVolumeSpecName: "config-data") pod "f320c18f-07fe-4255-a283-8128e0719f6a" (UID: "f320c18f-07fe-4255-a283-8128e0719f6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.882977 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.883094 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.883139 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.883156 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9twn6\" (UniqueName: \"kubernetes.io/projected/f320c18f-07fe-4255-a283-8128e0719f6a-kube-api-access-9twn6\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.883173 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:27 crc kubenswrapper[4726]: I1004 04:00:27.883188 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f320c18f-07fe-4255-a283-8128e0719f6a-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:28 crc kubenswrapper[4726]: E1004 04:00:28.075559 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:00:28 crc kubenswrapper[4726]: E1004 04:00:28.075774 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n94hc5h5h5b8h68dhbdh68fh55dh687h54h697h5d8h5bbh585h675h5c6h5cchdbhf4h565h96h65h5dbh55dhb9h66dh566h64dh5d9h589h59h694q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5mn4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7dbbf66849-ds92b_openstack(32f117f2-07b4-4f8b-9b51-63b1e7ffb97f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.621471 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ghnzv" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.717829 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ghnzv"] Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.723795 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ghnzv"] Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.826285 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sn774"] Oct 04 04:00:28 crc kubenswrapper[4726]: E1004 04:00:28.827428 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f320c18f-07fe-4255-a283-8128e0719f6a" containerName="keystone-bootstrap" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.827441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f320c18f-07fe-4255-a283-8128e0719f6a" containerName="keystone-bootstrap" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.828111 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f320c18f-07fe-4255-a283-8128e0719f6a" containerName="keystone-bootstrap" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.831962 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.840264 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2nwhj" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.840773 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.840988 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.841036 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:00:28 crc kubenswrapper[4726]: I1004 04:00:28.852091 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sn774"] Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.002513 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.002581 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.002680 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.002876 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.003072 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvkd\" (UniqueName: \"kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.003225 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.104115 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.104286 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2xgjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-72jth_openstack(1f5261e9-4ede-4c36-83cd-58cfe6176333): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105058 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvkd\" (UniqueName: \"kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105088 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105328 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105361 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.105405 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.106404 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-72jth" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.111949 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.112617 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.112857 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.125557 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.125604 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.131563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvkd\" (UniqueName: \"kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd\") pod \"keystone-bootstrap-sn774\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.155271 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.186807 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.196624 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309314 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309388 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309615 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9fl4\" (UniqueName: \"kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309655 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309763 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309850 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309879 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bvsz\" (UniqueName: \"kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.309974 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0\") pod \"ec74c29b-8285-4194-8d19-87475b516be3\" (UID: \"ec74c29b-8285-4194-8d19-87475b516be3\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.310005 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.310034 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run\") pod \"88d8101a-8928-43ec-9d26-4a9d1d01a669\" (UID: \"88d8101a-8928-43ec-9d26-4a9d1d01a669\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.310576 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs" (OuterVolumeSpecName: "logs") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.310937 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.313554 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts" (OuterVolumeSpecName: "scripts") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.314320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4" (OuterVolumeSpecName: "kube-api-access-q9fl4") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "kube-api-access-q9fl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.316499 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.331745 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz" (OuterVolumeSpecName: "kube-api-access-4bvsz") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "kube-api-access-4bvsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.336396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.358567 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.360996 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.362452 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config" (OuterVolumeSpecName: "config") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.369417 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data" (OuterVolumeSpecName: "config-data") pod "88d8101a-8928-43ec-9d26-4a9d1d01a669" (UID: "88d8101a-8928-43ec-9d26-4a9d1d01a669"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.378510 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.384913 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec74c29b-8285-4194-8d19-87475b516be3" (UID: "ec74c29b-8285-4194-8d19-87475b516be3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.412868 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.412910 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bvsz\" (UniqueName: \"kubernetes.io/projected/ec74c29b-8285-4194-8d19-87475b516be3-kube-api-access-4bvsz\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413166 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413322 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413344 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413491 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413511 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413524 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413535 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9fl4\" (UniqueName: \"kubernetes.io/projected/88d8101a-8928-43ec-9d26-4a9d1d01a669-kube-api-access-q9fl4\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413547 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d8101a-8928-43ec-9d26-4a9d1d01a669-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413558 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413570 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88d8101a-8928-43ec-9d26-4a9d1d01a669-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.413580 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec74c29b-8285-4194-8d19-87475b516be3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.430003 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.514830 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.592917 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.630645 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"88d8101a-8928-43ec-9d26-4a9d1d01a669","Type":"ContainerDied","Data":"2daea5937940635342fd2b19b15f3d2317b995fcd5702a73f63fcddefc577cfb"} Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.630721 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.632850 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.632844 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-4pcp4" event={"ID":"ec74c29b-8285-4194-8d19-87475b516be3","Type":"ContainerDied","Data":"c1aed412aa452a0d6bc3d97b9610c42a529e273aa9e8f737580793220ed4d138"} Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.635691 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-72jth" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.675813 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.684751 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-4pcp4"] Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.694219 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.701813 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.725958 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.726539 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="init" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726560 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="init" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.726582 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726588 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.726602 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-httpd" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726609 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-httpd" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.726636 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-log" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726643 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-log" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726807 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec74c29b-8285-4194-8d19-87475b516be3" containerName="dnsmasq-dns" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726819 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-log" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.726828 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" containerName="glance-httpd" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.727733 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.729841 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.733562 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.747797 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.752001 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.752187 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nhtwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-j9jwl_openstack(90894cee-ffab-4485-a221-703de154182f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:00:29 crc kubenswrapper[4726]: E1004 04:00:29.754093 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-j9jwl" podUID="90894cee-ffab-4485-a221-703de154182f" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.771768 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.785580 4726 scope.go:117] "RemoveContainer" containerID="8d6502612a8d96d2c27a59c733ba8d71d49002827af52e465691504bbfaadcce" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820283 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820397 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dczkr\" (UniqueName: \"kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820466 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820496 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820517 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.820540 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.861340 4726 scope.go:117] "RemoveContainer" containerID="235a4cd8da81c16a981fd44492ab3327ffb2e8e5a2475de19dea3237633099e0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.924891 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925272 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925470 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925548 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925628 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfgrc\" (UniqueName: \"kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925654 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925700 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts\") pod \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\" (UID: \"d7e592e7-e612-4c3f-912a-b15e701f0a4d\") " Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925942 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.925994 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926029 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926064 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926124 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs" (OuterVolumeSpecName: "logs") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926164 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926276 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dczkr\" (UniqueName: \"kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926317 4726 scope.go:117] "RemoveContainer" containerID="5c03e6e615995fc8526418938595ff2e3646f2be5b0d46459c0e18672d1acb2d" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.926422 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.927558 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.928759 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.929367 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.929381 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.932722 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.934454 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts" (OuterVolumeSpecName: "scripts") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.937281 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.940662 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc" (OuterVolumeSpecName: "kube-api-access-nfgrc") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "kube-api-access-nfgrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.950625 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.954350 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.960693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.977906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dczkr\" (UniqueName: \"kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:29 crc kubenswrapper[4726]: I1004 04:00:29.997840 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.014187 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.028290 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfgrc\" (UniqueName: \"kubernetes.io/projected/d7e592e7-e612-4c3f-912a-b15e701f0a4d-kube-api-access-nfgrc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.028318 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.028341 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.028352 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d7e592e7-e612-4c3f-912a-b15e701f0a4d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.028361 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.036607 4726 scope.go:117] "RemoveContainer" containerID="423b36e48db8be097e104487f0292e74540da76ede97b8b3cf9b07065e1e3b30" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.043049 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data" (OuterVolumeSpecName: "config-data") pod "d7e592e7-e612-4c3f-912a-b15e701f0a4d" (UID: "d7e592e7-e612-4c3f-912a-b15e701f0a4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.048562 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.051516 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.130649 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.130676 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e592e7-e612-4c3f-912a-b15e701f0a4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.302068 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.430200 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.440941 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.479395 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7786888bd-6khpf"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.521423 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d8101a-8928-43ec-9d26-4a9d1d01a669" path="/var/lib/kubelet/pods/88d8101a-8928-43ec-9d26-4a9d1d01a669/volumes" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.522902 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec74c29b-8285-4194-8d19-87475b516be3" path="/var/lib/kubelet/pods/ec74c29b-8285-4194-8d19-87475b516be3/volumes" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.524042 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f320c18f-07fe-4255-a283-8128e0719f6a" path="/var/lib/kubelet/pods/f320c18f-07fe-4255-a283-8128e0719f6a/volumes" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.551948 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sn774"] Oct 04 04:00:30 crc kubenswrapper[4726]: E1004 04:00:30.552247 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizon-7dbbf66849-ds92b" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" Oct 04 04:00:30 crc kubenswrapper[4726]: W1004 04:00:30.594239 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84f35fee_e3ff_49b7_980a_986aa0124565.slice/crio-4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71 WatchSource:0}: Error finding container 4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71: Status 404 returned error can't find the container with id 4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71 Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.647538 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerStarted","Data":"d32845fbbbd40ca9c1314f07f0f97f24428776b20c84b1f4666278a4c0beda45"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.664738 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.666757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" event={"ID":"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b","Type":"ContainerStarted","Data":"39de52332d3a940d5ea19a7ed218c4f0917a476b734bb9f40d0f0d03e67b61e7"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.671055 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerStarted","Data":"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.671084 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerStarted","Data":"2998495d4a268f42043a74303902308d460c37fb1a4240bc620df025b32b4335"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.672886 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7786888bd-6khpf" event={"ID":"888736e3-9e3d-402b-9ea3-79a0ba740fe7","Type":"ContainerStarted","Data":"02b614b235683987c226e6c67cdf7ea55e9c056815d91af72c260760235433fb"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.674594 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dbbf66849-ds92b" event={"ID":"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f","Type":"ContainerStarted","Data":"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.674670 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7dbbf66849-ds92b" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" containerName="horizon" containerID="cri-o://619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549" gracePeriod=30 Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.678953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sn774" event={"ID":"84f35fee-e3ff-49b7-980a-986aa0124565","Type":"ContainerStarted","Data":"4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.685714 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerStarted","Data":"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.695201 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d7e592e7-e612-4c3f-912a-b15e701f0a4d","Type":"ContainerDied","Data":"8af473e97dc9d3a6b72426f71b41c72d0a40519d6036fb0c43d14af57e402602"} Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.695259 4726 scope.go:117] "RemoveContainer" containerID="469e99b3af120fafd9dbe265bd5f91396f2a3c1ddac0a296db4f7f6729d5b633" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.695300 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.700071 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerStarted","Data":"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1"} Oct 04 04:00:30 crc kubenswrapper[4726]: E1004 04:00:30.704971 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-j9jwl" podUID="90894cee-ffab-4485-a221-703de154182f" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.755418 4726 scope.go:117] "RemoveContainer" containerID="5374d5402ff685151858fe00a7ce6245f7cf3ce18874401b7e2db3ba1f9c2b82" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.755774 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.768975 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.791047 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:00:30 crc kubenswrapper[4726]: E1004 04:00:30.791480 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-httpd" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.791492 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-httpd" Oct 04 04:00:30 crc kubenswrapper[4726]: E1004 04:00:30.791510 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-log" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.791516 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-log" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.791669 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-log" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.791683 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" containerName="glance-httpd" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.792613 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.795227 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.797813 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.799176 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.888893 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.951812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.951865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.951956 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjm9s\" (UniqueName: \"kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.951988 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.952059 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.952140 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.952181 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:30 crc kubenswrapper[4726]: I1004 04:00:30.952303 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053722 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053816 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053838 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053884 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjm9s\" (UniqueName: \"kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.053955 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.054148 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.054563 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.057074 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.059730 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.066557 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.067147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.067735 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.075853 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjm9s\" (UniqueName: \"kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.094520 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.159486 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.641003 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:00:31 crc kubenswrapper[4726]: W1004 04:00:31.681161 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadb80860_f5c3_417b_a55f_cc94f41568e3.slice/crio-41ebf9a89568f47326f5d12e1d7ce2d38b4283103aa570ca55f58c3c3c510afe WatchSource:0}: Error finding container 41ebf9a89568f47326f5d12e1d7ce2d38b4283103aa570ca55f58c3c3c510afe: Status 404 returned error can't find the container with id 41ebf9a89568f47326f5d12e1d7ce2d38b4283103aa570ca55f58c3c3c510afe Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.720432 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerStarted","Data":"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.720491 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cd8c86949-xgpbz" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon-log" containerID="cri-o://de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517" gracePeriod=30 Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.720569 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cd8c86949-xgpbz" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon" containerID="cri-o://91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7" gracePeriod=30 Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.732364 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerStarted","Data":"617f7095f19e49a0c5a39d938c529b27e0fb840bb9bce6a684f4ee9eac81ea5b"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.732413 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerStarted","Data":"a0d842fa13a2722bb75353ca431906776ad5260502a7c1254f23c450e66fcaf4"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.734911 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerStarted","Data":"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.734962 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerStarted","Data":"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.744486 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerStarted","Data":"41ebf9a89568f47326f5d12e1d7ce2d38b4283103aa570ca55f58c3c3c510afe"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.751515 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerStarted","Data":"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.751641 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7464dd7565-96l6d" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon-log" containerID="cri-o://29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1" gracePeriod=30 Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.751879 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7464dd7565-96l6d" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon" containerID="cri-o://7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed" gracePeriod=30 Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.768376 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5cd8c86949-xgpbz" podStartSLOduration=34.768359411 podStartE2EDuration="34.768359411s" podCreationTimestamp="2025-10-04 03:59:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:31.742145354 +0000 UTC m=+1205.916768587" watchObservedRunningTime="2025-10-04 04:00:31.768359411 +0000 UTC m=+1205.942982624" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.773211 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6b6cffb848-k9fxk" podStartSLOduration=28.773194271 podStartE2EDuration="28.773194271s" podCreationTimestamp="2025-10-04 04:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:31.768543396 +0000 UTC m=+1205.943166609" watchObservedRunningTime="2025-10-04 04:00:31.773194271 +0000 UTC m=+1205.947817484" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.776570 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7786888bd-6khpf" event={"ID":"888736e3-9e3d-402b-9ea3-79a0ba740fe7","Type":"ContainerStarted","Data":"c7ed7cfc314a8b58c79a6c7fb0262a68e193c23615d4dc81af7bcd63e39d01b2"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.776604 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7786888bd-6khpf" event={"ID":"888736e3-9e3d-402b-9ea3-79a0ba740fe7","Type":"ContainerStarted","Data":"ea65999c41682e10980457e0e674547bc8126eb09087daff79f1e69d68fcf4bb"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.779228 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sn774" event={"ID":"84f35fee-e3ff-49b7-980a-986aa0124565","Type":"ContainerStarted","Data":"dcfe4393bce42582a93d2a1065b96d65316474d253fdb4783fb43db8f0509c39"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.783061 4726 generic.go:334] "Generic (PLEG): container finished" podID="e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" containerID="b0251c200d8fb3d2d318b3cf1fdbe0df206b5904426ba80a9dad30b373391f77" exitCode=0 Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.783160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" event={"ID":"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b","Type":"ContainerDied","Data":"b0251c200d8fb3d2d318b3cf1fdbe0df206b5904426ba80a9dad30b373391f77"} Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.802405 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7464dd7565-96l6d" podStartSLOduration=3.911422592 podStartE2EDuration="40.802381878s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="2025-10-04 03:59:52.840183593 +0000 UTC m=+1167.014806806" lastFinishedPulling="2025-10-04 04:00:29.731142879 +0000 UTC m=+1203.905766092" observedRunningTime="2025-10-04 04:00:31.797424324 +0000 UTC m=+1205.972047537" watchObservedRunningTime="2025-10-04 04:00:31.802381878 +0000 UTC m=+1205.977005101" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.831398 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7786888bd-6khpf" podStartSLOduration=28.831383709 podStartE2EDuration="28.831383709s" podCreationTimestamp="2025-10-04 04:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:31.828591034 +0000 UTC m=+1206.003214247" watchObservedRunningTime="2025-10-04 04:00:31.831383709 +0000 UTC m=+1206.006006922" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.844020 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sn774" podStartSLOduration=3.844004759 podStartE2EDuration="3.844004759s" podCreationTimestamp="2025-10-04 04:00:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:31.843849025 +0000 UTC m=+1206.018472238" watchObservedRunningTime="2025-10-04 04:00:31.844004759 +0000 UTC m=+1206.018627972" Oct 04 04:00:31 crc kubenswrapper[4726]: I1004 04:00:31.975742 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7464dd7565-96l6d" Oct 04 04:00:32 crc kubenswrapper[4726]: I1004 04:00:32.479748 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 04:00:32 crc kubenswrapper[4726]: I1004 04:00:32.534128 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e592e7-e612-4c3f-912a-b15e701f0a4d" path="/var/lib/kubelet/pods/d7e592e7-e612-4c3f-912a-b15e701f0a4d/volumes" Oct 04 04:00:32 crc kubenswrapper[4726]: I1004 04:00:32.795577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerStarted","Data":"38b5892dcd0ab4e81a7fa3d1519031c715a6f637603d10a642538e3164101c96"} Oct 04 04:00:32 crc kubenswrapper[4726]: I1004 04:00:32.814764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerStarted","Data":"62218961f079497875f721f630e4e5747abbeb1de19e905266446070263c456b"} Oct 04 04:00:32 crc kubenswrapper[4726]: I1004 04:00:32.837555 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.837538115 podStartE2EDuration="3.837538115s" podCreationTimestamp="2025-10-04 04:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:32.833496056 +0000 UTC m=+1207.008119269" watchObservedRunningTime="2025-10-04 04:00:32.837538115 +0000 UTC m=+1207.012161318" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.149380 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.189290 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume\") pod \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.189420 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume\") pod \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.189456 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97xrt\" (UniqueName: \"kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt\") pod \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\" (UID: \"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b\") " Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.190162 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" (UID: "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.194826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt" (OuterVolumeSpecName: "kube-api-access-97xrt") pod "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" (UID: "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b"). InnerVolumeSpecName "kube-api-access-97xrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.211445 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" (UID: "e6d2aabc-dbc5-4442-af4e-2515c2d7e19b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.291551 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.291590 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97xrt\" (UniqueName: \"kubernetes.io/projected/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-kube-api-access-97xrt\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.291603 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.734146 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.734190 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.825840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerStarted","Data":"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe"} Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.828129 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" event={"ID":"e6d2aabc-dbc5-4442-af4e-2515c2d7e19b","Type":"ContainerDied","Data":"39de52332d3a940d5ea19a7ed218c4f0917a476b734bb9f40d0f0d03e67b61e7"} Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.828149 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39de52332d3a940d5ea19a7ed218c4f0917a476b734bb9f40d0f0d03e67b61e7" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.828206 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.831757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerStarted","Data":"c25f53e7f0cfa68844c99391f0d8fcd32ad1bf0f2e03f9af9e411c28e91932d6"} Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.835507 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:33 crc kubenswrapper[4726]: I1004 04:00:33.835543 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:00:34 crc kubenswrapper[4726]: I1004 04:00:34.891200 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.89118366 podStartE2EDuration="4.89118366s" podCreationTimestamp="2025-10-04 04:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:34.882506896 +0000 UTC m=+1209.057130109" watchObservedRunningTime="2025-10-04 04:00:34.89118366 +0000 UTC m=+1209.065806863" Oct 04 04:00:37 crc kubenswrapper[4726]: I1004 04:00:37.866193 4726 generic.go:334] "Generic (PLEG): container finished" podID="84f35fee-e3ff-49b7-980a-986aa0124565" containerID="dcfe4393bce42582a93d2a1065b96d65316474d253fdb4783fb43db8f0509c39" exitCode=0 Oct 04 04:00:37 crc kubenswrapper[4726]: I1004 04:00:37.866393 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sn774" event={"ID":"84f35fee-e3ff-49b7-980a-986aa0124565","Type":"ContainerDied","Data":"dcfe4393bce42582a93d2a1065b96d65316474d253fdb4783fb43db8f0509c39"} Oct 04 04:00:38 crc kubenswrapper[4726]: I1004 04:00:38.111218 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.049118 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.049378 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.089571 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.097249 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.894738 4726 generic.go:334] "Generic (PLEG): container finished" podID="ba7d46cf-02bd-4674-93e9-b339da1b568d" containerID="36ed94a606fbe6d95569c338bfc829a027d7fbb9bdf4294663ce49b3b96b58b0" exitCode=0 Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.896682 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q6284" event={"ID":"ba7d46cf-02bd-4674-93e9-b339da1b568d","Type":"ContainerDied","Data":"36ed94a606fbe6d95569c338bfc829a027d7fbb9bdf4294663ce49b3b96b58b0"} Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.896752 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:40 crc kubenswrapper[4726]: I1004 04:00:40.897547 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.028464 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.050906 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.050971 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.051187 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.051340 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.052011 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.052133 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqvkd\" (UniqueName: \"kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd\") pod \"84f35fee-e3ff-49b7-980a-986aa0124565\" (UID: \"84f35fee-e3ff-49b7-980a-986aa0124565\") " Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.057750 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts" (OuterVolumeSpecName: "scripts") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.062372 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd" (OuterVolumeSpecName: "kube-api-access-nqvkd") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "kube-api-access-nqvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.078233 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.084321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.103941 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data" (OuterVolumeSpecName: "config-data") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.125307 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84f35fee-e3ff-49b7-980a-986aa0124565" (UID: "84f35fee-e3ff-49b7-980a-986aa0124565"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155013 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155049 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155059 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155068 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqvkd\" (UniqueName: \"kubernetes.io/projected/84f35fee-e3ff-49b7-980a-986aa0124565-kube-api-access-nqvkd\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155078 4726 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.155086 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84f35fee-e3ff-49b7-980a-986aa0124565-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.159818 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.159874 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.196673 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.196757 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.905043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxqgf" event={"ID":"caf9ee75-66f7-4371-9adc-c0ad0249698c","Type":"ContainerStarted","Data":"13a6fc74ee64bc5598209bdc886cc65367098b521dea8cd60ae798ba13febe78"} Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.906891 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sn774" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.906905 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sn774" event={"ID":"84f35fee-e3ff-49b7-980a-986aa0124565","Type":"ContainerDied","Data":"4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71"} Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.906989 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b300def168a9a68e3320e82be7a8ff1419e855627e6d7068cbab0a247930a71" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.911430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerStarted","Data":"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4"} Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.911844 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.911889 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:00:41 crc kubenswrapper[4726]: I1004 04:00:41.943457 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xxqgf" podStartSLOduration=2.724453233 podStartE2EDuration="50.943438735s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="2025-10-04 03:59:53.430149672 +0000 UTC m=+1167.604772885" lastFinishedPulling="2025-10-04 04:00:41.649135184 +0000 UTC m=+1215.823758387" observedRunningTime="2025-10-04 04:00:41.938031329 +0000 UTC m=+1216.112654542" watchObservedRunningTime="2025-10-04 04:00:41.943438735 +0000 UTC m=+1216.118061948" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.250252 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-79b54b49b-mx2kb"] Oct 04 04:00:42 crc kubenswrapper[4726]: E1004 04:00:42.250979 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" containerName="collect-profiles" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.250993 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" containerName="collect-profiles" Oct 04 04:00:42 crc kubenswrapper[4726]: E1004 04:00:42.251030 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f35fee-e3ff-49b7-980a-986aa0124565" containerName="keystone-bootstrap" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.251037 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f35fee-e3ff-49b7-980a-986aa0124565" containerName="keystone-bootstrap" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.251222 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" containerName="collect-profiles" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.251238 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f35fee-e3ff-49b7-980a-986aa0124565" containerName="keystone-bootstrap" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.251806 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257176 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79b54b49b-mx2kb"] Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257389 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257565 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257608 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257668 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.257793 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2nwhj" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.259703 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-fernet-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-combined-ca-bundle\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380510 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-config-data\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380544 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-internal-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380601 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-credential-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380624 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-public-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380698 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snmc7\" (UniqueName: \"kubernetes.io/projected/cb8197e3-5f74-443a-bf22-78a28365cfaf-kube-api-access-snmc7\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.380731 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-scripts\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.400410 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q6284" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482070 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4b47\" (UniqueName: \"kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47\") pod \"ba7d46cf-02bd-4674-93e9-b339da1b568d\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482234 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config\") pod \"ba7d46cf-02bd-4674-93e9-b339da1b568d\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle\") pod \"ba7d46cf-02bd-4674-93e9-b339da1b568d\" (UID: \"ba7d46cf-02bd-4674-93e9-b339da1b568d\") " Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-fernet-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-combined-ca-bundle\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482604 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-config-data\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482630 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-internal-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482675 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-credential-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482693 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-public-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482736 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snmc7\" (UniqueName: \"kubernetes.io/projected/cb8197e3-5f74-443a-bf22-78a28365cfaf-kube-api-access-snmc7\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.482760 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-scripts\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.490494 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47" (OuterVolumeSpecName: "kube-api-access-c4b47") pod "ba7d46cf-02bd-4674-93e9-b339da1b568d" (UID: "ba7d46cf-02bd-4674-93e9-b339da1b568d"). InnerVolumeSpecName "kube-api-access-c4b47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.491609 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-config-data\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.493602 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-fernet-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.496765 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-combined-ca-bundle\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.497493 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-credential-keys\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.497642 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-internal-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.498632 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-public-tls-certs\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.501042 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8197e3-5f74-443a-bf22-78a28365cfaf-scripts\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.518990 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config" (OuterVolumeSpecName: "config") pod "ba7d46cf-02bd-4674-93e9-b339da1b568d" (UID: "ba7d46cf-02bd-4674-93e9-b339da1b568d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.519888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snmc7\" (UniqueName: \"kubernetes.io/projected/cb8197e3-5f74-443a-bf22-78a28365cfaf-kube-api-access-snmc7\") pod \"keystone-79b54b49b-mx2kb\" (UID: \"cb8197e3-5f74-443a-bf22-78a28365cfaf\") " pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.523222 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba7d46cf-02bd-4674-93e9-b339da1b568d" (UID: "ba7d46cf-02bd-4674-93e9-b339da1b568d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.584614 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.584645 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4b47\" (UniqueName: \"kubernetes.io/projected/ba7d46cf-02bd-4674-93e9-b339da1b568d-kube-api-access-c4b47\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.584656 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba7d46cf-02bd-4674-93e9-b339da1b568d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:42 crc kubenswrapper[4726]: I1004 04:00:42.702475 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.047337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9jwl" event={"ID":"90894cee-ffab-4485-a221-703de154182f","Type":"ContainerStarted","Data":"582be0e212516c976b60d9788c7eb4b00289ff683ab750e1b4481871333cf4a9"} Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.063666 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.063693 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.064320 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q6284" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.064406 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q6284" event={"ID":"ba7d46cf-02bd-4674-93e9-b339da1b568d","Type":"ContainerDied","Data":"b7c36f662d7e2525c2de8c24a7965ddcb8652c00c8c43b0eab2313f2c9249e39"} Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.064451 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7c36f662d7e2525c2de8c24a7965ddcb8652c00c8c43b0eab2313f2c9249e39" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.097567 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-j9jwl" podStartSLOduration=3.378656495 podStartE2EDuration="52.097548668s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="2025-10-04 03:59:53.247933082 +0000 UTC m=+1167.422556295" lastFinishedPulling="2025-10-04 04:00:41.966825255 +0000 UTC m=+1216.141448468" observedRunningTime="2025-10-04 04:00:43.097047585 +0000 UTC m=+1217.271670798" watchObservedRunningTime="2025-10-04 04:00:43.097548668 +0000 UTC m=+1217.272171881" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.266359 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79b54b49b-mx2kb"] Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.290256 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:43 crc kubenswrapper[4726]: E1004 04:00:43.290920 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba7d46cf-02bd-4674-93e9-b339da1b568d" containerName="neutron-db-sync" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.290934 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba7d46cf-02bd-4674-93e9-b339da1b568d" containerName="neutron-db-sync" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.291098 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba7d46cf-02bd-4674-93e9-b339da1b568d" containerName="neutron-db-sync" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.292045 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.323651 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424126 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424180 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424207 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9ws9\" (UniqueName: \"kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424227 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.424280 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.426865 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.429252 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.433751 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.434321 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.434449 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jgmht" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.434915 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.442372 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528859 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528909 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9ws9\" (UniqueName: \"kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528932 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528975 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.528999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529027 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529047 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529135 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxh7\" (UniqueName: \"kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529166 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529196 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.529906 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.530068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.530084 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.530471 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.555797 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9ws9\" (UniqueName: \"kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9\") pod \"dnsmasq-dns-5ccc5c4795-jk4lb\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.630660 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.630734 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.630771 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.630858 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxh7\" (UniqueName: \"kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.630890 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.636212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.636986 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.637226 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.641341 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.644669 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.690858 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxh7\" (UniqueName: \"kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7\") pod \"neutron-5dd6b788cb-dd5g5\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.740659 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.774737 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:43 crc kubenswrapper[4726]: I1004 04:00:43.848374 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7786888bd-6khpf" podUID="888736e3-9e3d-402b-9ea3-79a0ba740fe7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.089475 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.127776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79b54b49b-mx2kb" event={"ID":"cb8197e3-5f74-443a-bf22-78a28365cfaf","Type":"ContainerStarted","Data":"8268a98a1ac5259c1795af85b598789d50ec6ac61e17f96eb77f744d83ddb2e5"} Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.128086 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79b54b49b-mx2kb" event={"ID":"cb8197e3-5f74-443a-bf22-78a28365cfaf","Type":"ContainerStarted","Data":"e4a23ff781a88007c8852da2a236511fd7a2084cfbcfa4857d3eb926ea347961"} Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.128311 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.226863 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.226951 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.239495 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.253578 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-79b54b49b-mx2kb" podStartSLOduration=2.253564472 podStartE2EDuration="2.253564472s" podCreationTimestamp="2025-10-04 04:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:44.182486536 +0000 UTC m=+1218.357109749" watchObservedRunningTime="2025-10-04 04:00:44.253564472 +0000 UTC m=+1218.428187685" Oct 04 04:00:44 crc kubenswrapper[4726]: I1004 04:00:44.344667 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.147086 4726 generic.go:334] "Generic (PLEG): container finished" podID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerID="87aac1c62cfe24c6655f333a5213811b86045bb7462e061f66871e4b4c1bcb44" exitCode=0 Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.147886 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" event={"ID":"e15942aa-cc35-421a-961b-56cb2a99f6ce","Type":"ContainerDied","Data":"87aac1c62cfe24c6655f333a5213811b86045bb7462e061f66871e4b4c1bcb44"} Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.147978 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" event={"ID":"e15942aa-cc35-421a-961b-56cb2a99f6ce","Type":"ContainerStarted","Data":"f08862bc74e5acbde9df8b409bdf1ddfccd450da00910290cd8fcb053dea4f83"} Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.199118 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-72jth" event={"ID":"1f5261e9-4ede-4c36-83cd-58cfe6176333","Type":"ContainerStarted","Data":"e5c4b24a358c073895352df74618c137ab92ae661ae669c963644d144ab8cd7f"} Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.221433 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerStarted","Data":"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604"} Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.226425 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerStarted","Data":"d02f490eed6321c0045320e6144dcfa0884cd5fbb1d6bede490fd17e184af498"} Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.317263 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-72jth" podStartSLOduration=4.169873707 podStartE2EDuration="54.317245878s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="2025-10-04 03:59:52.910692213 +0000 UTC m=+1167.085315426" lastFinishedPulling="2025-10-04 04:00:43.058064384 +0000 UTC m=+1217.232687597" observedRunningTime="2025-10-04 04:00:45.237223451 +0000 UTC m=+1219.411846664" watchObservedRunningTime="2025-10-04 04:00:45.317245878 +0000 UTC m=+1219.491869091" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.650511 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5bd64b8469-btvtm"] Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.652126 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.658373 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.658427 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.683944 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5bd64b8469-btvtm"] Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.808998 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-internal-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809069 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-public-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809096 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-ovndb-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809129 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-combined-ca-bundle\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809149 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809181 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb2qm\" (UniqueName: \"kubernetes.io/projected/d97c6241-e072-4012-a61e-c5e855407e99-kube-api-access-bb2qm\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.809278 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-httpd-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.911017 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-combined-ca-bundle\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.911076 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.911139 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb2qm\" (UniqueName: \"kubernetes.io/projected/d97c6241-e072-4012-a61e-c5e855407e99-kube-api-access-bb2qm\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.911176 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-httpd-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.914492 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-internal-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.914864 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-public-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.914905 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-ovndb-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.919720 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-combined-ca-bundle\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.924607 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-internal-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.924670 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-public-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.928670 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-httpd-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.932464 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-ovndb-tls-certs\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.935860 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d97c6241-e072-4012-a61e-c5e855407e99-config\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.946706 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb2qm\" (UniqueName: \"kubernetes.io/projected/d97c6241-e072-4012-a61e-c5e855407e99-kube-api-access-bb2qm\") pod \"neutron-5bd64b8469-btvtm\" (UID: \"d97c6241-e072-4012-a61e-c5e855407e99\") " pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:45 crc kubenswrapper[4726]: I1004 04:00:45.991930 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.248243 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" event={"ID":"e15942aa-cc35-421a-961b-56cb2a99f6ce","Type":"ContainerStarted","Data":"169ce95ffefb46af608204f6f29dfe9ca255dc52b6a06314e063b9a8e3871687"} Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.248401 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.258827 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerStarted","Data":"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f"} Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.259365 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.275648 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" podStartSLOduration=3.275632376 podStartE2EDuration="3.275632376s" podCreationTimestamp="2025-10-04 04:00:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:46.271295179 +0000 UTC m=+1220.445918392" watchObservedRunningTime="2025-10-04 04:00:46.275632376 +0000 UTC m=+1220.450255589" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.299357 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5dd6b788cb-dd5g5" podStartSLOduration=3.299323425 podStartE2EDuration="3.299323425s" podCreationTimestamp="2025-10-04 04:00:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:46.288699838 +0000 UTC m=+1220.463323051" watchObservedRunningTime="2025-10-04 04:00:46.299323425 +0000 UTC m=+1220.473946638" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.812150 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5bd64b8469-btvtm"] Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.829682 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.829807 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:00:46 crc kubenswrapper[4726]: I1004 04:00:46.843332 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.305662 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5bd64b8469-btvtm" event={"ID":"d97c6241-e072-4012-a61e-c5e855407e99","Type":"ContainerStarted","Data":"8a71da90058ef46cf691319cf0ec6fa3e75f1154e2de686dfdb2425ffc6eac94"} Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.305988 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5bd64b8469-btvtm" event={"ID":"d97c6241-e072-4012-a61e-c5e855407e99","Type":"ContainerStarted","Data":"b05ae24f0cfea37f8170498b4a9be41f1d8c53a7a98d6c99f68a381b8f7e1ede"} Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.313502 4726 generic.go:334] "Generic (PLEG): container finished" podID="caf9ee75-66f7-4371-9adc-c0ad0249698c" containerID="13a6fc74ee64bc5598209bdc886cc65367098b521dea8cd60ae798ba13febe78" exitCode=0 Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.313583 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxqgf" event={"ID":"caf9ee75-66f7-4371-9adc-c0ad0249698c","Type":"ContainerDied","Data":"13a6fc74ee64bc5598209bdc886cc65367098b521dea8cd60ae798ba13febe78"} Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.340200 4726 generic.go:334] "Generic (PLEG): container finished" podID="90894cee-ffab-4485-a221-703de154182f" containerID="582be0e212516c976b60d9788c7eb4b00289ff683ab750e1b4481871333cf4a9" exitCode=0 Oct 04 04:00:47 crc kubenswrapper[4726]: I1004 04:00:47.340339 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9jwl" event={"ID":"90894cee-ffab-4485-a221-703de154182f","Type":"ContainerDied","Data":"582be0e212516c976b60d9788c7eb4b00289ff683ab750e1b4481871333cf4a9"} Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.358434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5bd64b8469-btvtm" event={"ID":"d97c6241-e072-4012-a61e-c5e855407e99","Type":"ContainerStarted","Data":"ce517daa4b02c271c49e97b4acaaf2075a824420a93a1bdf640287e67d46ed2c"} Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.380450 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5bd64b8469-btvtm" podStartSLOduration=3.380432609 podStartE2EDuration="3.380432609s" podCreationTimestamp="2025-10-04 04:00:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:48.377981243 +0000 UTC m=+1222.552604466" watchObservedRunningTime="2025-10-04 04:00:48.380432609 +0000 UTC m=+1222.555055822" Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.802149 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9jwl" Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.894588 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxqgf" Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.914468 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle\") pod \"90894cee-ffab-4485-a221-703de154182f\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.914565 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhtwc\" (UniqueName: \"kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc\") pod \"90894cee-ffab-4485-a221-703de154182f\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.914647 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data\") pod \"90894cee-ffab-4485-a221-703de154182f\" (UID: \"90894cee-ffab-4485-a221-703de154182f\") " Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.923944 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "90894cee-ffab-4485-a221-703de154182f" (UID: "90894cee-ffab-4485-a221-703de154182f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.927851 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc" (OuterVolumeSpecName: "kube-api-access-nhtwc") pod "90894cee-ffab-4485-a221-703de154182f" (UID: "90894cee-ffab-4485-a221-703de154182f"). InnerVolumeSpecName "kube-api-access-nhtwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:48 crc kubenswrapper[4726]: I1004 04:00:48.973993 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90894cee-ffab-4485-a221-703de154182f" (UID: "90894cee-ffab-4485-a221-703de154182f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015503 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs\") pod \"caf9ee75-66f7-4371-9adc-c0ad0249698c\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015763 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxtlh\" (UniqueName: \"kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh\") pod \"caf9ee75-66f7-4371-9adc-c0ad0249698c\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015828 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle\") pod \"caf9ee75-66f7-4371-9adc-c0ad0249698c\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015863 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs" (OuterVolumeSpecName: "logs") pod "caf9ee75-66f7-4371-9adc-c0ad0249698c" (UID: "caf9ee75-66f7-4371-9adc-c0ad0249698c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015889 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts\") pod \"caf9ee75-66f7-4371-9adc-c0ad0249698c\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.015999 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data\") pod \"caf9ee75-66f7-4371-9adc-c0ad0249698c\" (UID: \"caf9ee75-66f7-4371-9adc-c0ad0249698c\") " Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.016415 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.016432 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90894cee-ffab-4485-a221-703de154182f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.016441 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/caf9ee75-66f7-4371-9adc-c0ad0249698c-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.016451 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhtwc\" (UniqueName: \"kubernetes.io/projected/90894cee-ffab-4485-a221-703de154182f-kube-api-access-nhtwc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.021387 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts" (OuterVolumeSpecName: "scripts") pod "caf9ee75-66f7-4371-9adc-c0ad0249698c" (UID: "caf9ee75-66f7-4371-9adc-c0ad0249698c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.033820 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh" (OuterVolumeSpecName: "kube-api-access-gxtlh") pod "caf9ee75-66f7-4371-9adc-c0ad0249698c" (UID: "caf9ee75-66f7-4371-9adc-c0ad0249698c"). InnerVolumeSpecName "kube-api-access-gxtlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.042860 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data" (OuterVolumeSpecName: "config-data") pod "caf9ee75-66f7-4371-9adc-c0ad0249698c" (UID: "caf9ee75-66f7-4371-9adc-c0ad0249698c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.050833 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caf9ee75-66f7-4371-9adc-c0ad0249698c" (UID: "caf9ee75-66f7-4371-9adc-c0ad0249698c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.117785 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.117832 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.117844 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf9ee75-66f7-4371-9adc-c0ad0249698c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.117857 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxtlh\" (UniqueName: \"kubernetes.io/projected/caf9ee75-66f7-4371-9adc-c0ad0249698c-kube-api-access-gxtlh\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.370007 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9jwl" event={"ID":"90894cee-ffab-4485-a221-703de154182f","Type":"ContainerDied","Data":"ec0b089df2ef265f090cf72d8935f68ece32401e2e9efec525ce7ff8f5d318ed"} Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.370048 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec0b089df2ef265f090cf72d8935f68ece32401e2e9efec525ce7ff8f5d318ed" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.370138 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9jwl" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.372092 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xxqgf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.372779 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xxqgf" event={"ID":"caf9ee75-66f7-4371-9adc-c0ad0249698c","Type":"ContainerDied","Data":"cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93"} Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.372813 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd9f21deb9e6c49702427ac9d6948ecb477515b57d89b87272c9f25dc1626d93" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.372828 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.430778 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6f697b85fb-mcpgk"] Oct 04 04:00:49 crc kubenswrapper[4726]: E1004 04:00:49.435239 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90894cee-ffab-4485-a221-703de154182f" containerName="barbican-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.435269 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="90894cee-ffab-4485-a221-703de154182f" containerName="barbican-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: E1004 04:00:49.435299 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" containerName="placement-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.435308 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" containerName="placement-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.435638 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" containerName="placement-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.435664 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="90894cee-ffab-4485-a221-703de154182f" containerName="barbican-db-sync" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.436586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.439007 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.440417 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.440584 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-xs8d7" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.440748 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.441326 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.456032 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f697b85fb-mcpgk"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530514 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-scripts\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530577 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5md8\" (UniqueName: \"kubernetes.io/projected/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-kube-api-access-z5md8\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-config-data\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530664 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-public-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530724 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-combined-ca-bundle\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530753 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-internal-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.530889 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-logs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.612186 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5fccfff78b-bs7hd"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.614074 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.618760 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fccfff78b-bs7hd"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.621485 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mxjcb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.621531 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.621719 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634471 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-scripts\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5md8\" (UniqueName: \"kubernetes.io/projected/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-kube-api-access-z5md8\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634588 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-config-data\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634649 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-public-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634689 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-combined-ca-bundle\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634727 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-internal-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.634820 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-logs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.635593 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-logs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.656194 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-public-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.656248 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-78dd78b9bc-pwdzf"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.656515 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-internal-tls-certs\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.657639 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.660039 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-scripts\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.660860 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-config-data\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.663519 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dd78b9bc-pwdzf"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.670343 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-combined-ca-bundle\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.678073 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.692470 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5md8\" (UniqueName: \"kubernetes.io/projected/c10188ee-0fc9-4985-a21c-88fbc85e1e3c-kube-api-access-z5md8\") pod \"placement-6f697b85fb-mcpgk\" (UID: \"c10188ee-0fc9-4985-a21c-88fbc85e1e3c\") " pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737760 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737797 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d885c-08ac-43eb-a1a7-2829719d4187-logs\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data-custom\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737877 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mr88\" (UniqueName: \"kubernetes.io/projected/0f684dad-f165-4648-a5b0-e1e94b2cdc44-kube-api-access-7mr88\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737912 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-combined-ca-bundle\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737943 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data-custom\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.737980 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f684dad-f165-4648-a5b0-e1e94b2cdc44-logs\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.738001 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-combined-ca-bundle\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.738023 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccsjm\" (UniqueName: \"kubernetes.io/projected/be9d885c-08ac-43eb-a1a7-2829719d4187-kube-api-access-ccsjm\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.738049 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.763006 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.772152 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.772375 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="dnsmasq-dns" containerID="cri-o://169ce95ffefb46af608204f6f29dfe9ca255dc52b6a06314e063b9a8e3871687" gracePeriod=10 Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.837275 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.839029 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851013 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-combined-ca-bundle\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851123 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data-custom\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851207 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f684dad-f165-4648-a5b0-e1e94b2cdc44-logs\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851243 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-combined-ca-bundle\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccsjm\" (UniqueName: \"kubernetes.io/projected/be9d885c-08ac-43eb-a1a7-2829719d4187-kube-api-access-ccsjm\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851330 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851390 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851409 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d885c-08ac-43eb-a1a7-2829719d4187-logs\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851448 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data-custom\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.851559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mr88\" (UniqueName: \"kubernetes.io/projected/0f684dad-f165-4648-a5b0-e1e94b2cdc44-kube-api-access-7mr88\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.852156 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f684dad-f165-4648-a5b0-e1e94b2cdc44-logs\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.852555 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d885c-08ac-43eb-a1a7-2829719d4187-logs\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.889782 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-combined-ca-bundle\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.890804 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.891212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-config-data-custom\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.891690 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f684dad-f165-4648-a5b0-e1e94b2cdc44-combined-ca-bundle\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.892805 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data-custom\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.893927 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d885c-08ac-43eb-a1a7-2829719d4187-config-data\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.894333 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.907661 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccsjm\" (UniqueName: \"kubernetes.io/projected/be9d885c-08ac-43eb-a1a7-2829719d4187-kube-api-access-ccsjm\") pod \"barbican-worker-78dd78b9bc-pwdzf\" (UID: \"be9d885c-08ac-43eb-a1a7-2829719d4187\") " pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.920285 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.922822 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.928579 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.929725 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mr88\" (UniqueName: \"kubernetes.io/projected/0f684dad-f165-4648-a5b0-e1e94b2cdc44-kube-api-access-7mr88\") pod \"barbican-keystone-listener-5fccfff78b-bs7hd\" (UID: \"0f684dad-f165-4648-a5b0-e1e94b2cdc44\") " pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.946989 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969450 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969481 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969510 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969555 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl667\" (UniqueName: \"kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:49 crc kubenswrapper[4726]: I1004 04:00:49.969660 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.015798 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.040497 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071093 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl667\" (UniqueName: \"kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071173 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvs2\" (UniqueName: \"kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071232 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071278 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071322 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071342 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071453 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.071472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.072423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.072794 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.073050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.073131 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.080770 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.108824 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl667\" (UniqueName: \"kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667\") pod \"dnsmasq-dns-688c87cc99-nxtpb\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.189215 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvs2\" (UniqueName: \"kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.189345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.189389 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.189412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.189441 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.191586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.201453 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.205621 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.216520 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.216664 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f697b85fb-mcpgk"] Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.235760 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvs2\" (UniqueName: \"kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.293721 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom\") pod \"barbican-api-57f84cf794-p9968\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.293933 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.400937 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f697b85fb-mcpgk" event={"ID":"c10188ee-0fc9-4985-a21c-88fbc85e1e3c","Type":"ContainerStarted","Data":"33ccff78ea65ebc5a7d966cf0dda109de62a9e1746fcb9a5d1153ce26a74c637"} Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.412896 4726 generic.go:334] "Generic (PLEG): container finished" podID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerID="169ce95ffefb46af608204f6f29dfe9ca255dc52b6a06314e063b9a8e3871687" exitCode=0 Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.414271 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" event={"ID":"e15942aa-cc35-421a-961b-56cb2a99f6ce","Type":"ContainerDied","Data":"169ce95ffefb46af608204f6f29dfe9ca255dc52b6a06314e063b9a8e3871687"} Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.739449 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807334 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807500 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807527 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9ws9\" (UniqueName: \"kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807556 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.807609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config\") pod \"e15942aa-cc35-421a-961b-56cb2a99f6ce\" (UID: \"e15942aa-cc35-421a-961b-56cb2a99f6ce\") " Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.808424 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5fccfff78b-bs7hd"] Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.827133 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9" (OuterVolumeSpecName: "kube-api-access-v9ws9") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "kube-api-access-v9ws9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.902514 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.914033 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9ws9\" (UniqueName: \"kubernetes.io/projected/e15942aa-cc35-421a-961b-56cb2a99f6ce-kube-api-access-v9ws9\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.914062 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.916406 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.920236 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.963556 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config" (OuterVolumeSpecName: "config") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.983755 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e15942aa-cc35-421a-961b-56cb2a99f6ce" (UID: "e15942aa-cc35-421a-961b-56cb2a99f6ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:00:50 crc kubenswrapper[4726]: I1004 04:00:50.996000 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dd78b9bc-pwdzf"] Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.023145 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.023170 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.023179 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.023188 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e15942aa-cc35-421a-961b-56cb2a99f6ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.103169 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.125585 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.434652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" event={"ID":"be9d885c-08ac-43eb-a1a7-2829719d4187","Type":"ContainerStarted","Data":"fc8c45038445e264788ee21590ba0b087afbcc2189fce6da1e03631481a9e584"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.457086 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f697b85fb-mcpgk" event={"ID":"c10188ee-0fc9-4985-a21c-88fbc85e1e3c","Type":"ContainerStarted","Data":"a024ccce89cceee0de8c03a7ea506b781ccdc3dc41f60c35b4fead7cc887aff2"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.457140 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f697b85fb-mcpgk" event={"ID":"c10188ee-0fc9-4985-a21c-88fbc85e1e3c","Type":"ContainerStarted","Data":"037d4e151b8f5d23d094462c6bf09a15a65516ec37e1caeba96fecc65d62c6e4"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.458828 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.458863 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.493311 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerStarted","Data":"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.493353 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerStarted","Data":"6cb3717b92169681c638b60ea0920de69fbd1adc4ce2eb38ee6109b09d1be81c"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.494533 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6f697b85fb-mcpgk" podStartSLOduration=2.494506792 podStartE2EDuration="2.494506792s" podCreationTimestamp="2025-10-04 04:00:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:51.479542199 +0000 UTC m=+1225.654165412" watchObservedRunningTime="2025-10-04 04:00:51.494506792 +0000 UTC m=+1225.669130005" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.498788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" event={"ID":"0f684dad-f165-4648-a5b0-e1e94b2cdc44","Type":"ContainerStarted","Data":"a9e33cada6866975a44c284fadd25e458519b3c0841633a74bbaba6fd4d84d98"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.506352 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" event={"ID":"e15942aa-cc35-421a-961b-56cb2a99f6ce","Type":"ContainerDied","Data":"f08862bc74e5acbde9df8b409bdf1ddfccd450da00910290cd8fcb053dea4f83"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.506399 4726 scope.go:117] "RemoveContainer" containerID="169ce95ffefb46af608204f6f29dfe9ca255dc52b6a06314e063b9a8e3871687" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.506506 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-jk4lb" Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.517975 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerStarted","Data":"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.518270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerStarted","Data":"1b8ef4a2661a0d81c8484d9532e4bdc6f7900f61342750a536febba58fe42933"} Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.561231 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.566767 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-jk4lb"] Oct 04 04:00:51 crc kubenswrapper[4726]: I1004 04:00:51.573342 4726 scope.go:117] "RemoveContainer" containerID="87aac1c62cfe24c6655f333a5213811b86045bb7462e061f66871e4b4c1bcb44" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.515057 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" path="/var/lib/kubelet/pods/e15942aa-cc35-421a-961b-56cb2a99f6ce/volumes" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.536859 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerStarted","Data":"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55"} Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.538953 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.538997 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.561520 4726 generic.go:334] "Generic (PLEG): container finished" podID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerID="5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4" exitCode=0 Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.561582 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerDied","Data":"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4"} Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.635588 4726 generic.go:334] "Generic (PLEG): container finished" podID="1f5261e9-4ede-4c36-83cd-58cfe6176333" containerID="e5c4b24a358c073895352df74618c137ab92ae661ae669c963644d144ab8cd7f" exitCode=0 Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.636338 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-72jth" event={"ID":"1f5261e9-4ede-4c36-83cd-58cfe6176333","Type":"ContainerDied","Data":"e5c4b24a358c073895352df74618c137ab92ae661ae669c963644d144ab8cd7f"} Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.647998 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-56bb546-6dhls"] Oct 04 04:00:52 crc kubenswrapper[4726]: E1004 04:00:52.648627 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="dnsmasq-dns" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.648644 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="dnsmasq-dns" Oct 04 04:00:52 crc kubenswrapper[4726]: E1004 04:00:52.648676 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="init" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.648684 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="init" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.648861 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e15942aa-cc35-421a-961b-56cb2a99f6ce" containerName="dnsmasq-dns" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.676589 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.733088 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.735592 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.780613 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57f84cf794-p9968" podStartSLOduration=3.780592602 podStartE2EDuration="3.780592602s" podCreationTimestamp="2025-10-04 04:00:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:00:52.606033288 +0000 UTC m=+1226.780656501" watchObservedRunningTime="2025-10-04 04:00:52.780592602 +0000 UTC m=+1226.955215815" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.810622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.810735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-combined-ca-bundle\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.810786 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74c5864-48ad-4376-b318-071c1925ad4c-logs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.810852 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data-custom\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.810942 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q72l\" (UniqueName: \"kubernetes.io/projected/d74c5864-48ad-4376-b318-071c1925ad4c-kube-api-access-4q72l\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.811033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-public-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.811163 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-internal-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.826058 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56bb546-6dhls"] Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916021 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q72l\" (UniqueName: \"kubernetes.io/projected/d74c5864-48ad-4376-b318-071c1925ad4c-kube-api-access-4q72l\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916101 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-public-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-internal-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916210 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916236 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-combined-ca-bundle\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916259 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74c5864-48ad-4376-b318-071c1925ad4c-logs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916286 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data-custom\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.916820 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d74c5864-48ad-4376-b318-071c1925ad4c-logs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.919942 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-internal-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.920363 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-public-tls-certs\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.923375 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-combined-ca-bundle\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.924558 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.924667 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d74c5864-48ad-4376-b318-071c1925ad4c-config-data-custom\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:52 crc kubenswrapper[4726]: I1004 04:00:52.930148 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q72l\" (UniqueName: \"kubernetes.io/projected/d74c5864-48ad-4376-b318-071c1925ad4c-kube-api-access-4q72l\") pod \"barbican-api-56bb546-6dhls\" (UID: \"d74c5864-48ad-4376-b318-071c1925ad4c\") " pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:53 crc kubenswrapper[4726]: I1004 04:00:53.122315 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:00:53 crc kubenswrapper[4726]: I1004 04:00:53.734749 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:00:53 crc kubenswrapper[4726]: I1004 04:00:53.835792 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7786888bd-6khpf" podUID="888736e3-9e3d-402b-9ea3-79a0ba740fe7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 04 04:00:57 crc kubenswrapper[4726]: I1004 04:00:57.797381 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-57f84cf794-p9968" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.228495 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-72jth" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.341983 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.342560 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.342611 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.342662 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xgjg\" (UniqueName: \"kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.342817 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.342838 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data\") pod \"1f5261e9-4ede-4c36-83cd-58cfe6176333\" (UID: \"1f5261e9-4ede-4c36-83cd-58cfe6176333\") " Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.343482 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.350010 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.353448 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg" (OuterVolumeSpecName: "kube-api-access-2xgjg") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "kube-api-access-2xgjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.360374 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts" (OuterVolumeSpecName: "scripts") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.404544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.419124 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data" (OuterVolumeSpecName: "config-data") pod "1f5261e9-4ede-4c36-83cd-58cfe6176333" (UID: "1f5261e9-4ede-4c36-83cd-58cfe6176333"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445296 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445329 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445341 4726 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445354 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5261e9-4ede-4c36-83cd-58cfe6176333-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445362 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1f5261e9-4ede-4c36-83cd-58cfe6176333-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.445371 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xgjg\" (UniqueName: \"kubernetes.io/projected/1f5261e9-4ede-4c36-83cd-58cfe6176333-kube-api-access-2xgjg\") on node \"crc\" DevicePath \"\"" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.710842 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-72jth" event={"ID":"1f5261e9-4ede-4c36-83cd-58cfe6176333","Type":"ContainerDied","Data":"702795c02ae3ca02fcc6158cadaf5017e7822629eee9bae32a3d54a06c1ac5a7"} Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.710890 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="702795c02ae3ca02fcc6158cadaf5017e7822629eee9bae32a3d54a06c1ac5a7" Oct 04 04:00:58 crc kubenswrapper[4726]: I1004 04:00:58.710901 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-72jth" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.334225 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-56bb546-6dhls"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.566525 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.574098 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:00:59 crc kubenswrapper[4726]: E1004 04:00:59.574480 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" containerName="cinder-db-sync" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.574492 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" containerName="cinder-db-sync" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.574676 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" containerName="cinder-db-sync" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.575586 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.593091 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.593755 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.593908 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.594567 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dkv48" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.594689 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.605615 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.607210 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.671189 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.782511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56bb546-6dhls" event={"ID":"d74c5864-48ad-4376-b318-071c1925ad4c","Type":"ContainerStarted","Data":"2d92fd73db73738c5a6038023595b1006deb2c5d20d0f5c261aed5205fab819e"} Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.784727 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.787281 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.789439 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.792924 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794463 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfk9j\" (UniqueName: \"kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794522 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794545 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794569 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794616 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794695 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794727 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794843 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794876 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.794944 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl4ck\" (UniqueName: \"kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.795183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896355 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896409 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896439 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfk9j\" (UniqueName: \"kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896457 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896473 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896496 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896522 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896578 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn9sl\" (UniqueName: \"kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896599 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896618 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896670 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896687 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896720 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896746 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896761 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896782 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896797 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.896815 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl4ck\" (UniqueName: \"kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.897625 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.898394 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.898466 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.898671 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.898766 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.901038 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.905887 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.910096 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.912500 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.921229 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfk9j\" (UniqueName: \"kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j\") pod \"dnsmasq-dns-6bb4fc677f-mh5qd\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.926465 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl4ck\" (UniqueName: \"kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.926842 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data\") pod \"cinder-scheduler-0\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " pod="openstack/cinder-scheduler-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998099 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998161 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998215 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998303 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn9sl\" (UniqueName: \"kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998345 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998373 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:00:59 crc kubenswrapper[4726]: I1004 04:00:59.998391 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.001853 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.002206 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.004987 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.008390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.011639 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.012399 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.020592 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn9sl\" (UniqueName: \"kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl\") pod \"cinder-api-0\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.138766 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.146156 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325841-fnfsx"] Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.147359 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.161505 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325841-fnfsx"] Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.187273 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.201568 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.201640 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.201715 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbz86\" (UniqueName: \"kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.201740 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.214139 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.303376 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbz86\" (UniqueName: \"kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.303418 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.303516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.303560 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.312892 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.312967 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.316959 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.333169 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbz86\" (UniqueName: \"kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86\") pod \"keystone-cron-29325841-fnfsx\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.403828 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.795136 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.802377 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" event={"ID":"be9d885c-08ac-43eb-a1a7-2829719d4187","Type":"ContainerStarted","Data":"a95e8be96ffb38daefc039aa1346d488b760bc4ef51a9f8b8d42072c93dd3120"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.802416 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" event={"ID":"be9d885c-08ac-43eb-a1a7-2829719d4187","Type":"ContainerStarted","Data":"ab15a0be99f4d77ec33c36ddf9156e212cf4e98ccadddbf11dec9a03301d07eb"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.802429 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.818735 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerStarted","Data":"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.818947 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-central-agent" containerID="cri-o://2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5" gracePeriod=30 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.819274 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.819306 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="sg-core" containerID="cri-o://575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4" gracePeriod=30 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.819287 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="proxy-httpd" containerID="cri-o://4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6" gracePeriod=30 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.819385 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-notification-agent" containerID="cri-o://7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe" gracePeriod=30 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.832845 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerStarted","Data":"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.833016 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="dnsmasq-dns" containerID="cri-o://129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23" gracePeriod=10 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.833093 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.835578 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-78dd78b9bc-pwdzf" podStartSLOduration=3.359443246 podStartE2EDuration="11.835555285s" podCreationTimestamp="2025-10-04 04:00:49 +0000 UTC" firstStartedPulling="2025-10-04 04:00:51.005839174 +0000 UTC m=+1225.180462387" lastFinishedPulling="2025-10-04 04:00:59.481951213 +0000 UTC m=+1233.656574426" observedRunningTime="2025-10-04 04:01:00.825251728 +0000 UTC m=+1234.999874941" watchObservedRunningTime="2025-10-04 04:01:00.835555285 +0000 UTC m=+1235.010178498" Oct 04 04:01:00 crc kubenswrapper[4726]: W1004 04:01:00.836233 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod626168ca_767c_43ae_8836_b2c75a8a7286.slice/crio-cbdae37efbee918a8fd065050bfaa7fa8abc27268187a638abd0b0370aa3d9b7 WatchSource:0}: Error finding container cbdae37efbee918a8fd065050bfaa7fa8abc27268187a638abd0b0370aa3d9b7: Status 404 returned error can't find the container with id cbdae37efbee918a8fd065050bfaa7fa8abc27268187a638abd0b0370aa3d9b7 Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.843982 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" event={"ID":"0f684dad-f165-4648-a5b0-e1e94b2cdc44","Type":"ContainerStarted","Data":"8ebeb9c6c9af70382ad831cb4707f0c1a93d69181303741796272509956995c0"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.844039 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" event={"ID":"0f684dad-f165-4648-a5b0-e1e94b2cdc44","Type":"ContainerStarted","Data":"949afabbd23fd6a2dfb66c392aa888c4ba3e08bddbaa48aaaa10ffefb3408f41"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.857973 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56bb546-6dhls" event={"ID":"d74c5864-48ad-4376-b318-071c1925ad4c","Type":"ContainerStarted","Data":"be8434d9962401884ff930d1c97351a5bf52d82f69e1c77e3797b8053960d5ee"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.858014 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-56bb546-6dhls" event={"ID":"d74c5864-48ad-4376-b318-071c1925ad4c","Type":"ContainerStarted","Data":"8981b72ecbc07aa673319c7e4804655abf325476c04acefc8b84148f03cb25d4"} Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.858359 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.858421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.880797 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.202375567 podStartE2EDuration="1m9.880780073s" podCreationTimestamp="2025-10-04 03:59:51 +0000 UTC" firstStartedPulling="2025-10-04 03:59:53.045854796 +0000 UTC m=+1167.220478009" lastFinishedPulling="2025-10-04 04:00:59.724259302 +0000 UTC m=+1233.898882515" observedRunningTime="2025-10-04 04:01:00.851471168 +0000 UTC m=+1235.026094381" watchObservedRunningTime="2025-10-04 04:01:00.880780073 +0000 UTC m=+1235.055403286" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.904596 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" podStartSLOduration=11.904575464 podStartE2EDuration="11.904575464s" podCreationTimestamp="2025-10-04 04:00:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:00.87782077 +0000 UTC m=+1235.052443983" watchObservedRunningTime="2025-10-04 04:01:00.904575464 +0000 UTC m=+1235.079198677" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.918311 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5fccfff78b-bs7hd" podStartSLOduration=3.289107921 podStartE2EDuration="11.918292206s" podCreationTimestamp="2025-10-04 04:00:49 +0000 UTC" firstStartedPulling="2025-10-04 04:00:50.852898032 +0000 UTC m=+1225.027521245" lastFinishedPulling="2025-10-04 04:00:59.482082317 +0000 UTC m=+1233.656705530" observedRunningTime="2025-10-04 04:01:00.89615317 +0000 UTC m=+1235.070776383" watchObservedRunningTime="2025-10-04 04:01:00.918292206 +0000 UTC m=+1235.092915429" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.939061 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-56bb546-6dhls" podStartSLOduration=8.939042223 podStartE2EDuration="8.939042223s" podCreationTimestamp="2025-10-04 04:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:00.921041902 +0000 UTC m=+1235.095665115" watchObservedRunningTime="2025-10-04 04:01:00.939042223 +0000 UTC m=+1235.113665436" Oct 04 04:01:00 crc kubenswrapper[4726]: I1004 04:01:00.963772 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.063191 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325841-fnfsx"] Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.471433 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.631390 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.639668 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mn4c\" (UniqueName: \"kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c\") pod \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.639927 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs\") pod \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.639990 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data\") pod \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.640016 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key\") pod \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.640039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts\") pod \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\" (UID: \"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.640638 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs" (OuterVolumeSpecName: "logs") pod "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" (UID: "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.654346 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c" (OuterVolumeSpecName: "kube-api-access-5mn4c") pod "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" (UID: "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f"). InnerVolumeSpecName "kube-api-access-5mn4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.667220 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" (UID: "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.717943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts" (OuterVolumeSpecName: "scripts") pod "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" (UID: "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748181 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl667\" (UniqueName: \"kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748270 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748286 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748301 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748335 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748375 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config\") pod \"47179504-09c2-4acd-a84d-9e1c3510b2fd\" (UID: \"47179504-09c2-4acd-a84d-9e1c3510b2fd\") " Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748853 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748869 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748893 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mn4c\" (UniqueName: \"kubernetes.io/projected/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-kube-api-access-5mn4c\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.748903 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.758720 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667" (OuterVolumeSpecName: "kube-api-access-wl667") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "kube-api-access-wl667". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:01 crc kubenswrapper[4726]: I1004 04:01:01.764833 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data" (OuterVolumeSpecName: "config-data") pod "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" (UID: "32f117f2-07b4-4f8b-9b51-63b1e7ffb97f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.853345 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl667\" (UniqueName: \"kubernetes.io/projected/47179504-09c2-4acd-a84d-9e1c3510b2fd-kube-api-access-wl667\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.853372 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.862901 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.879329 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.885941 4726 generic.go:334] "Generic (PLEG): container finished" podID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerID="cdf750fb434c1778bfd98585f593db00d740410b7f80d2188619ea27d7425865" exitCode=0 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.886016 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" event={"ID":"5f5f1547-f888-4cc7-97f3-84ce9117f5c8","Type":"ContainerDied","Data":"cdf750fb434c1778bfd98585f593db00d740410b7f80d2188619ea27d7425865"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.886047 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" event={"ID":"5f5f1547-f888-4cc7-97f3-84ce9117f5c8","Type":"ContainerStarted","Data":"fa8fa0908805788706aa6427e09c2fe9525b3d1d0946b3b128fec06d9336efdb"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.896047 4726 generic.go:334] "Generic (PLEG): container finished" podID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerID="de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517" exitCode=137 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.896496 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerDied","Data":"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.900411 4726 generic.go:334] "Generic (PLEG): container finished" podID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerID="29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1" exitCode=137 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.900452 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerDied","Data":"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.901600 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config" (OuterVolumeSpecName: "config") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.902414 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerStarted","Data":"f9905f43699dc37319892cbb0b7d72b9f9b8d5422a6f9673b0919ca02a129508"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.905544 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerStarted","Data":"cbdae37efbee918a8fd065050bfaa7fa8abc27268187a638abd0b0370aa3d9b7"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.908020 4726 generic.go:334] "Generic (PLEG): container finished" podID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerID="129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23" exitCode=0 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.908060 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerDied","Data":"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.908079 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" event={"ID":"47179504-09c2-4acd-a84d-9e1c3510b2fd","Type":"ContainerDied","Data":"6cb3717b92169681c638b60ea0920de69fbd1adc4ce2eb38ee6109b09d1be81c"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.908095 4726 scope.go:117] "RemoveContainer" containerID="129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.908326 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-nxtpb" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910665 4726 generic.go:334] "Generic (PLEG): container finished" podID="67531077-c83d-49e5-b077-87d3671159f1" containerID="4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6" exitCode=0 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910688 4726 generic.go:334] "Generic (PLEG): container finished" podID="67531077-c83d-49e5-b077-87d3671159f1" containerID="575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4" exitCode=2 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910697 4726 generic.go:334] "Generic (PLEG): container finished" podID="67531077-c83d-49e5-b077-87d3671159f1" containerID="2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5" exitCode=0 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910738 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerDied","Data":"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerDied","Data":"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.910780 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerDied","Data":"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.912323 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325841-fnfsx" event={"ID":"db67cb41-86ca-41f3-b012-1bfc33e84a37","Type":"ContainerStarted","Data":"6fa8edc31ae4bc2989f6afa627e5acc0f29b1e03b31a1f88aaa42dcdf58baec9"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.912370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325841-fnfsx" event={"ID":"db67cb41-86ca-41f3-b012-1bfc33e84a37","Type":"ContainerStarted","Data":"35aa2930ad7ff623d92c0d8785c158473d6e3b1ff9ba8f2626c5cbd822e3570f"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.916258 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.916360 4726 generic.go:334] "Generic (PLEG): container finished" podID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" containerID="619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549" exitCode=137 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.916459 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dbbf66849-ds92b" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.916820 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dbbf66849-ds92b" event={"ID":"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f","Type":"ContainerDied","Data":"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.916874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dbbf66849-ds92b" event={"ID":"32f117f2-07b4-4f8b-9b51-63b1e7ffb97f","Type":"ContainerDied","Data":"fc20dc8ee4a30c26dc877fdacdc6c5d7c635114885f3f643c580f666177ea268"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.943472 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.952788 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325841-fnfsx" podStartSLOduration=1.952765913 podStartE2EDuration="1.952765913s" podCreationTimestamp="2025-10-04 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:01.93253386 +0000 UTC m=+1236.107157073" watchObservedRunningTime="2025-10-04 04:01:01.952765913 +0000 UTC m=+1236.127389126" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.955371 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.955397 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.955409 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.955422 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:01.974951 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47179504-09c2-4acd-a84d-9e1c3510b2fd" (UID: "47179504-09c2-4acd-a84d-9e1c3510b2fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.056808 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47179504-09c2-4acd-a84d-9e1c3510b2fd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.393344 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.590393 4726 scope.go:117] "RemoveContainer" containerID="5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.637320 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7464dd7565-96l6d" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.691669 4726 scope.go:117] "RemoveContainer" containerID="129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23" Oct 04 04:01:02 crc kubenswrapper[4726]: E1004 04:01:02.699652 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23\": container with ID starting with 129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23 not found: ID does not exist" containerID="129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.699691 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23"} err="failed to get container status \"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23\": rpc error: code = NotFound desc = could not find container \"129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23\": container with ID starting with 129eefaae98c13669a443685b7641ae8367972f4ecd8f06212d5d0cf2fbf2d23 not found: ID does not exist" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.699712 4726 scope.go:117] "RemoveContainer" containerID="5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.701164 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:01:02 crc kubenswrapper[4726]: E1004 04:01:02.710374 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4\": container with ID starting with 5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4 not found: ID does not exist" containerID="5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.710418 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4"} err="failed to get container status \"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4\": rpc error: code = NotFound desc = could not find container \"5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4\": container with ID starting with 5883fa9fbd7690033bfdb274ae544cbbeaaa3f742fca20e2f1cd06aa1f7b9cb4 not found: ID does not exist" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.710450 4726 scope.go:117] "RemoveContainer" containerID="619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.712870 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.720025 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-nxtpb"] Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.746754 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.756051 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7dbbf66849-ds92b"] Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.770606 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvgjj\" (UniqueName: \"kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj\") pod \"71dec834-afb0-4ace-98b5-3be330ae5bd5\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.770653 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts\") pod \"71dec834-afb0-4ace-98b5-3be330ae5bd5\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.770688 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key\") pod \"71dec834-afb0-4ace-98b5-3be330ae5bd5\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.770726 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data\") pod \"71dec834-afb0-4ace-98b5-3be330ae5bd5\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.770760 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs\") pod \"71dec834-afb0-4ace-98b5-3be330ae5bd5\" (UID: \"71dec834-afb0-4ace-98b5-3be330ae5bd5\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.771538 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs" (OuterVolumeSpecName: "logs") pod "71dec834-afb0-4ace-98b5-3be330ae5bd5" (UID: "71dec834-afb0-4ace-98b5-3be330ae5bd5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.776279 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj" (OuterVolumeSpecName: "kube-api-access-zvgjj") pod "71dec834-afb0-4ace-98b5-3be330ae5bd5" (UID: "71dec834-afb0-4ace-98b5-3be330ae5bd5"). InnerVolumeSpecName "kube-api-access-zvgjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.777209 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "71dec834-afb0-4ace-98b5-3be330ae5bd5" (UID: "71dec834-afb0-4ace-98b5-3be330ae5bd5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.801798 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts" (OuterVolumeSpecName: "scripts") pod "71dec834-afb0-4ace-98b5-3be330ae5bd5" (UID: "71dec834-afb0-4ace-98b5-3be330ae5bd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.801850 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data" (OuterVolumeSpecName: "config-data") pod "71dec834-afb0-4ace-98b5-3be330ae5bd5" (UID: "71dec834-afb0-4ace-98b5-3be330ae5bd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.802758 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.873626 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts\") pod \"20a0920f-7366-4d50-96e5-212cb1c30cf1\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.873687 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data\") pod \"20a0920f-7366-4d50-96e5-212cb1c30cf1\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.873781 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mns7z\" (UniqueName: \"kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z\") pod \"20a0920f-7366-4d50-96e5-212cb1c30cf1\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.873813 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key\") pod \"20a0920f-7366-4d50-96e5-212cb1c30cf1\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.877884 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs\") pod \"20a0920f-7366-4d50-96e5-212cb1c30cf1\" (UID: \"20a0920f-7366-4d50-96e5-212cb1c30cf1\") " Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878396 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "20a0920f-7366-4d50-96e5-212cb1c30cf1" (UID: "20a0920f-7366-4d50-96e5-212cb1c30cf1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878542 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvgjj\" (UniqueName: \"kubernetes.io/projected/71dec834-afb0-4ace-98b5-3be330ae5bd5-kube-api-access-zvgjj\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878556 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878565 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/71dec834-afb0-4ace-98b5-3be330ae5bd5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878573 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/71dec834-afb0-4ace-98b5-3be330ae5bd5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878581 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71dec834-afb0-4ace-98b5-3be330ae5bd5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878590 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/20a0920f-7366-4d50-96e5-212cb1c30cf1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.878952 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs" (OuterVolumeSpecName: "logs") pod "20a0920f-7366-4d50-96e5-212cb1c30cf1" (UID: "20a0920f-7366-4d50-96e5-212cb1c30cf1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.883422 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z" (OuterVolumeSpecName: "kube-api-access-mns7z") pod "20a0920f-7366-4d50-96e5-212cb1c30cf1" (UID: "20a0920f-7366-4d50-96e5-212cb1c30cf1"). InnerVolumeSpecName "kube-api-access-mns7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.898869 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data" (OuterVolumeSpecName: "config-data") pod "20a0920f-7366-4d50-96e5-212cb1c30cf1" (UID: "20a0920f-7366-4d50-96e5-212cb1c30cf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.911540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts" (OuterVolumeSpecName: "scripts") pod "20a0920f-7366-4d50-96e5-212cb1c30cf1" (UID: "20a0920f-7366-4d50-96e5-212cb1c30cf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.949817 4726 generic.go:334] "Generic (PLEG): container finished" podID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerID="7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed" exitCode=137 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.949873 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerDied","Data":"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.949899 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7464dd7565-96l6d" event={"ID":"71dec834-afb0-4ace-98b5-3be330ae5bd5","Type":"ContainerDied","Data":"26b166ddfd1608e2b858f55b473e617de5ef444a811c56694a7ab9e7dde290c3"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.949954 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7464dd7565-96l6d" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.956182 4726 generic.go:334] "Generic (PLEG): container finished" podID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerID="91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7" exitCode=137 Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.956260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerDied","Data":"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.956298 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cd8c86949-xgpbz" event={"ID":"20a0920f-7366-4d50-96e5-212cb1c30cf1","Type":"ContainerDied","Data":"2998495d4a268f42043a74303902308d460c37fb1a4240bc620df025b32b4335"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.956389 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cd8c86949-xgpbz" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.962026 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerStarted","Data":"3e007c7d33ea562866b4c155cd180e78a46ffca18f99b994448dece04dcb1fa9"} Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.979929 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20a0920f-7366-4d50-96e5-212cb1c30cf1-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.979949 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.979958 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/20a0920f-7366-4d50-96e5-212cb1c30cf1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.979967 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mns7z\" (UniqueName: \"kubernetes.io/projected/20a0920f-7366-4d50-96e5-212cb1c30cf1-kube-api-access-mns7z\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:02 crc kubenswrapper[4726]: I1004 04:01:02.999441 4726 scope.go:117] "RemoveContainer" containerID="619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549" Oct 04 04:01:03 crc kubenswrapper[4726]: E1004 04:01:03.006122 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549\": container with ID starting with 619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549 not found: ID does not exist" containerID="619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.006155 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549"} err="failed to get container status \"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549\": rpc error: code = NotFound desc = could not find container \"619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549\": container with ID starting with 619278ba696c50fb4293ccd73b654f4b9b2671d343959b0c97032a4b1b99e549 not found: ID does not exist" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.006181 4726 scope.go:117] "RemoveContainer" containerID="7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.016999 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.032507 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7464dd7565-96l6d"] Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.045914 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.053979 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5cd8c86949-xgpbz"] Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.224547 4726 scope.go:117] "RemoveContainer" containerID="29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.392389 4726 scope.go:117] "RemoveContainer" containerID="7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed" Oct 04 04:01:03 crc kubenswrapper[4726]: E1004 04:01:03.392957 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed\": container with ID starting with 7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed not found: ID does not exist" containerID="7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.392998 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed"} err="failed to get container status \"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed\": rpc error: code = NotFound desc = could not find container \"7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed\": container with ID starting with 7886286d5383dbc17553756da6befad3201fa41f6c0a445fcf334f962a196fed not found: ID does not exist" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.393026 4726 scope.go:117] "RemoveContainer" containerID="29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1" Oct 04 04:01:03 crc kubenswrapper[4726]: E1004 04:01:03.393392 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1\": container with ID starting with 29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1 not found: ID does not exist" containerID="29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.393444 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1"} err="failed to get container status \"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1\": rpc error: code = NotFound desc = could not find container \"29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1\": container with ID starting with 29f307dc9738de6ef23d6ad3bcf7824b227d3012b369f8250a64c99c91212ab1 not found: ID does not exist" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.393477 4726 scope.go:117] "RemoveContainer" containerID="91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.625142 4726 scope.go:117] "RemoveContainer" containerID="de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.657645 4726 scope.go:117] "RemoveContainer" containerID="91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7" Oct 04 04:01:03 crc kubenswrapper[4726]: E1004 04:01:03.658019 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7\": container with ID starting with 91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7 not found: ID does not exist" containerID="91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.658060 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7"} err="failed to get container status \"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7\": rpc error: code = NotFound desc = could not find container \"91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7\": container with ID starting with 91d0755b8ecd4256ad791190a0ef356ca6ef024426e51ddcff63d0af768e52c7 not found: ID does not exist" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.658086 4726 scope.go:117] "RemoveContainer" containerID="de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517" Oct 04 04:01:03 crc kubenswrapper[4726]: E1004 04:01:03.658388 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517\": container with ID starting with de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517 not found: ID does not exist" containerID="de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.658418 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517"} err="failed to get container status \"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517\": rpc error: code = NotFound desc = could not find container \"de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517\": container with ID starting with de5b8a8fc6c009387c09f669998bc017ea775ed861c7a22786145ae61e59a517 not found: ID does not exist" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.983474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerStarted","Data":"8c0e952b0cc7d7d9d048d68777b98e3a8be638f7322a6a9498f03f2d64e13826"} Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.983648 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api-log" containerID="cri-o://3e007c7d33ea562866b4c155cd180e78a46ffca18f99b994448dece04dcb1fa9" gracePeriod=30 Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.983696 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api" containerID="cri-o://8c0e952b0cc7d7d9d048d68777b98e3a8be638f7322a6a9498f03f2d64e13826" gracePeriod=30 Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.983873 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.989434 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerStarted","Data":"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83"} Oct 04 04:01:03 crc kubenswrapper[4726]: I1004 04:01:03.989498 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerStarted","Data":"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45"} Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.000278 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" event={"ID":"5f5f1547-f888-4cc7-97f3-84ce9117f5c8","Type":"ContainerStarted","Data":"36c16ca90803489bd416b0a75f9350c121f6743ad6e3bc03208b982406944302"} Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.000355 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.004259 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.00424187 podStartE2EDuration="5.00424187s" podCreationTimestamp="2025-10-04 04:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:04.001778882 +0000 UTC m=+1238.176402145" watchObservedRunningTime="2025-10-04 04:01:04.00424187 +0000 UTC m=+1238.178865093" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.028650 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" podStartSLOduration=5.028629728 podStartE2EDuration="5.028629728s" podCreationTimestamp="2025-10-04 04:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:04.026921871 +0000 UTC m=+1238.201545094" watchObservedRunningTime="2025-10-04 04:01:04.028629728 +0000 UTC m=+1238.203252951" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.520723 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" path="/var/lib/kubelet/pods/20a0920f-7366-4d50-96e5-212cb1c30cf1/volumes" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.522059 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" path="/var/lib/kubelet/pods/32f117f2-07b4-4f8b-9b51-63b1e7ffb97f/volumes" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.523076 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" path="/var/lib/kubelet/pods/47179504-09c2-4acd-a84d-9e1c3510b2fd/volumes" Oct 04 04:01:04 crc kubenswrapper[4726]: I1004 04:01:04.525404 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" path="/var/lib/kubelet/pods/71dec834-afb0-4ace-98b5-3be330ae5bd5/volumes" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.019432 4726 generic.go:334] "Generic (PLEG): container finished" podID="db67cb41-86ca-41f3-b012-1bfc33e84a37" containerID="6fa8edc31ae4bc2989f6afa627e5acc0f29b1e03b31a1f88aaa42dcdf58baec9" exitCode=0 Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.019584 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325841-fnfsx" event={"ID":"db67cb41-86ca-41f3-b012-1bfc33e84a37","Type":"ContainerDied","Data":"6fa8edc31ae4bc2989f6afa627e5acc0f29b1e03b31a1f88aaa42dcdf58baec9"} Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.022088 4726 generic.go:334] "Generic (PLEG): container finished" podID="f039778e-109d-4d80-a839-be2db809794d" containerID="8c0e952b0cc7d7d9d048d68777b98e3a8be638f7322a6a9498f03f2d64e13826" exitCode=0 Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.022161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerDied","Data":"8c0e952b0cc7d7d9d048d68777b98e3a8be638f7322a6a9498f03f2d64e13826"} Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.022215 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerDied","Data":"3e007c7d33ea562866b4c155cd180e78a46ffca18f99b994448dece04dcb1fa9"} Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.022172 4726 generic.go:334] "Generic (PLEG): container finished" podID="f039778e-109d-4d80-a839-be2db809794d" containerID="3e007c7d33ea562866b4c155cd180e78a46ffca18f99b994448dece04dcb1fa9" exitCode=143 Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.073288 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.8071477 podStartE2EDuration="6.073264518s" podCreationTimestamp="2025-10-04 04:00:59 +0000 UTC" firstStartedPulling="2025-10-04 04:01:00.839171405 +0000 UTC m=+1235.013794628" lastFinishedPulling="2025-10-04 04:01:02.105288233 +0000 UTC m=+1236.279911446" observedRunningTime="2025-10-04 04:01:05.0657967 +0000 UTC m=+1239.240420023" watchObservedRunningTime="2025-10-04 04:01:05.073264518 +0000 UTC m=+1239.247887771" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.140202 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.728709 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.875756 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.875841 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.876047 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.876324 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn9sl\" (UniqueName: \"kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.876379 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.876412 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.876687 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs\") pod \"f039778e-109d-4d80-a839-be2db809794d\" (UID: \"f039778e-109d-4d80-a839-be2db809794d\") " Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.877901 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs" (OuterVolumeSpecName: "logs") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.877917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.905292 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts" (OuterVolumeSpecName: "scripts") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.905423 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl" (OuterVolumeSpecName: "kube-api-access-wn9sl") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "kube-api-access-wn9sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.922441 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.922925 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.940334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data" (OuterVolumeSpecName: "config-data") pod "f039778e-109d-4d80-a839-be2db809794d" (UID: "f039778e-109d-4d80-a839-be2db809794d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980486 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn9sl\" (UniqueName: \"kubernetes.io/projected/f039778e-109d-4d80-a839-be2db809794d-kube-api-access-wn9sl\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980706 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980717 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980726 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f039778e-109d-4d80-a839-be2db809794d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980734 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f039778e-109d-4d80-a839-be2db809794d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980742 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.980750 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f039778e-109d-4d80-a839-be2db809794d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:05 crc kubenswrapper[4726]: I1004 04:01:05.985467 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.047691 4726 generic.go:334] "Generic (PLEG): container finished" podID="67531077-c83d-49e5-b077-87d3671159f1" containerID="7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe" exitCode=0 Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.047756 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerDied","Data":"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe"} Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.047786 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67531077-c83d-49e5-b077-87d3671159f1","Type":"ContainerDied","Data":"153d89033e70faf5f09403c4376fb6737982e2e6d55a22c3feaf1ee6bc8a32ca"} Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.047808 4726 scope.go:117] "RemoveContainer" containerID="4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.047974 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.062813 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.063366 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f039778e-109d-4d80-a839-be2db809794d","Type":"ContainerDied","Data":"f9905f43699dc37319892cbb0b7d72b9f9b8d5422a6f9673b0919ca02a129508"} Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.107588 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.123727 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.133749 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134083 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="sg-core" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134099 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="sg-core" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134115 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134122 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134128 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="proxy-httpd" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134134 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="proxy-httpd" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134240 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134249 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134262 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134269 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134300 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-notification-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134307 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-notification-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134318 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134323 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134340 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134345 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134361 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134369 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134383 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="dnsmasq-dns" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134390 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="dnsmasq-dns" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134404 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-central-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134411 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-central-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134420 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="init" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134427 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="init" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.134440 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134445 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134614 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="sg-core" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134624 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134631 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134642 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134651 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-notification-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134660 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f117f2-07b4-4f8b-9b51-63b1e7ffb97f" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134674 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="47179504-09c2-4acd-a84d-9e1c3510b2fd" containerName="dnsmasq-dns" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134688 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dec834-afb0-4ace-98b5-3be330ae5bd5" containerName="horizon-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134695 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a0920f-7366-4d50-96e5-212cb1c30cf1" containerName="horizon" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134703 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f039778e-109d-4d80-a839-be2db809794d" containerName="cinder-api-log" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134714 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="ceilometer-central-agent" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.134722 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="67531077-c83d-49e5-b077-87d3671159f1" containerName="proxy-httpd" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.135618 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.153782 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.154043 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.154217 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.182195 4726 scope.go:117] "RemoveContainer" containerID="575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.185274 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.185787 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.185862 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss25f\" (UniqueName: \"kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.185937 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.185970 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.186039 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.186140 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.186178 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle\") pod \"67531077-c83d-49e5-b077-87d3671159f1\" (UID: \"67531077-c83d-49e5-b077-87d3671159f1\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.187728 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.190830 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.194248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts" (OuterVolumeSpecName: "scripts") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.206547 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f" (OuterVolumeSpecName: "kube-api-access-ss25f") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "kube-api-access-ss25f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.221290 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.223622 4726 scope.go:117] "RemoveContainer" containerID="7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.279718 4726 scope.go:117] "RemoveContainer" containerID="2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289209 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-scripts\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289302 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289382 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc91ace6-aab5-4a5c-9549-131d3303f5d5-logs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289425 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289704 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwsv9\" (UniqueName: \"kubernetes.io/projected/cc91ace6-aab5-4a5c-9549-131d3303f5d5-kube-api-access-jwsv9\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289750 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289820 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc91ace6-aab5-4a5c-9549-131d3303f5d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289941 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289954 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss25f\" (UniqueName: \"kubernetes.io/projected/67531077-c83d-49e5-b077-87d3671159f1-kube-api-access-ss25f\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289965 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67531077-c83d-49e5-b077-87d3671159f1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289974 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.289983 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.292292 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.302017 4726 scope.go:117] "RemoveContainer" containerID="4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.303281 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6\": container with ID starting with 4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6 not found: ID does not exist" containerID="4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.303317 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6"} err="failed to get container status \"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6\": rpc error: code = NotFound desc = could not find container \"4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6\": container with ID starting with 4be2402fbcceaabda38778232110a0a1ecbc3b74595f96ae209579f8e760abc6 not found: ID does not exist" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.303344 4726 scope.go:117] "RemoveContainer" containerID="575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.316360 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4\": container with ID starting with 575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4 not found: ID does not exist" containerID="575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.316402 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4"} err="failed to get container status \"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4\": rpc error: code = NotFound desc = could not find container \"575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4\": container with ID starting with 575d2735f387866d3565da3e81c12ebb204621ae61c55f4edc0c1eb9a0de68a4 not found: ID does not exist" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.316432 4726 scope.go:117] "RemoveContainer" containerID="7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.316884 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe\": container with ID starting with 7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe not found: ID does not exist" containerID="7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.316920 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe"} err="failed to get container status \"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe\": rpc error: code = NotFound desc = could not find container \"7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe\": container with ID starting with 7a2250da8d978016c5c67c78add8cfacb83d7735196099685e05315a0fc40ffe not found: ID does not exist" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.316946 4726 scope.go:117] "RemoveContainer" containerID="2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5" Oct 04 04:01:06 crc kubenswrapper[4726]: E1004 04:01:06.317181 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5\": container with ID starting with 2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5 not found: ID does not exist" containerID="2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.317225 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5"} err="failed to get container status \"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5\": rpc error: code = NotFound desc = could not find container \"2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5\": container with ID starting with 2937c1921692a4e2ebd9e6b72423952e60d12afb2b19280f5b75b82ec85a9ef5 not found: ID does not exist" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.317238 4726 scope.go:117] "RemoveContainer" containerID="8c0e952b0cc7d7d9d048d68777b98e3a8be638f7322a6a9498f03f2d64e13826" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.323282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data" (OuterVolumeSpecName: "config-data") pod "67531077-c83d-49e5-b077-87d3671159f1" (UID: "67531077-c83d-49e5-b077-87d3671159f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.351001 4726 scope.go:117] "RemoveContainer" containerID="3e007c7d33ea562866b4c155cd180e78a46ffca18f99b994448dece04dcb1fa9" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391547 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwsv9\" (UniqueName: \"kubernetes.io/projected/cc91ace6-aab5-4a5c-9549-131d3303f5d5-kube-api-access-jwsv9\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391585 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391619 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc91ace6-aab5-4a5c-9549-131d3303f5d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391643 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-scripts\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391678 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391723 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391740 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc91ace6-aab5-4a5c-9549-131d3303f5d5-logs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391759 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391775 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391817 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.391827 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67531077-c83d-49e5-b077-87d3671159f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.392647 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.395828 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc91ace6-aab5-4a5c-9549-131d3303f5d5-logs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.397242 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc91ace6-aab5-4a5c-9549-131d3303f5d5-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.405903 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data-custom\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.405955 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.407843 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.409411 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-scripts\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.411253 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.413640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwsv9\" (UniqueName: \"kubernetes.io/projected/cc91ace6-aab5-4a5c-9549-131d3303f5d5-kube-api-access-jwsv9\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.414219 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-config-data\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.415867 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc91ace6-aab5-4a5c-9549-131d3303f5d5-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cc91ace6-aab5-4a5c-9549-131d3303f5d5\") " pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.417251 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.419564 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.422081 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.422309 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.425882 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.446492 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.485563 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.525154 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67531077-c83d-49e5-b077-87d3671159f1" path="/var/lib/kubelet/pods/67531077-c83d-49e5-b077-87d3671159f1/volumes" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.526381 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f039778e-109d-4d80-a839-be2db809794d" path="/var/lib/kubelet/pods/f039778e-109d-4d80-a839-be2db809794d/volumes" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.598837 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data\") pod \"db67cb41-86ca-41f3-b012-1bfc33e84a37\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599006 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbz86\" (UniqueName: \"kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86\") pod \"db67cb41-86ca-41f3-b012-1bfc33e84a37\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599047 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys\") pod \"db67cb41-86ca-41f3-b012-1bfc33e84a37\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599110 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle\") pod \"db67cb41-86ca-41f3-b012-1bfc33e84a37\" (UID: \"db67cb41-86ca-41f3-b012-1bfc33e84a37\") " Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599325 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599362 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599386 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599433 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599451 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599470 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc9k8\" (UniqueName: \"kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.599566 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.608391 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86" (OuterVolumeSpecName: "kube-api-access-vbz86") pod "db67cb41-86ca-41f3-b012-1bfc33e84a37" (UID: "db67cb41-86ca-41f3-b012-1bfc33e84a37"). InnerVolumeSpecName "kube-api-access-vbz86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.608410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "db67cb41-86ca-41f3-b012-1bfc33e84a37" (UID: "db67cb41-86ca-41f3-b012-1bfc33e84a37"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.640298 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db67cb41-86ca-41f3-b012-1bfc33e84a37" (UID: "db67cb41-86ca-41f3-b012-1bfc33e84a37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737232 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737258 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737295 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737336 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737351 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc9k8\" (UniqueName: \"kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737452 4726 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737466 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.737479 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbz86\" (UniqueName: \"kubernetes.io/projected/db67cb41-86ca-41f3-b012-1bfc33e84a37-kube-api-access-vbz86\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.738449 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.739723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.739865 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data" (OuterVolumeSpecName: "config-data") pod "db67cb41-86ca-41f3-b012-1bfc33e84a37" (UID: "db67cb41-86ca-41f3-b012-1bfc33e84a37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.741018 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.745937 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.746182 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.752743 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.759774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc9k8\" (UniqueName: \"kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8\") pod \"ceilometer-0\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.772976 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.798447 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.839690 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db67cb41-86ca-41f3-b012-1bfc33e84a37-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:06 crc kubenswrapper[4726]: I1004 04:01:06.907860 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.030177 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.101470 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325841-fnfsx" Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.101525 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325841-fnfsx" event={"ID":"db67cb41-86ca-41f3-b012-1bfc33e84a37","Type":"ContainerDied","Data":"35aa2930ad7ff623d92c0d8785c158473d6e3b1ff9ba8f2626c5cbd822e3570f"} Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.101570 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35aa2930ad7ff623d92c0d8785c158473d6e3b1ff9ba8f2626c5cbd822e3570f" Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.106138 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc91ace6-aab5-4a5c-9549-131d3303f5d5","Type":"ContainerStarted","Data":"d0fa88f15fc0d6e6b1cb42c0ce770017b900f516ca6d1c938d36e55ec9708468"} Oct 04 04:01:07 crc kubenswrapper[4726]: I1004 04:01:07.358067 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:07 crc kubenswrapper[4726]: W1004 04:01:07.361655 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71eeaf87_9cec_4ba0_9702_99bacffc46a3.slice/crio-224863ed2625caf14a09031edc6fc0dda69fdeb66afe26f4d73ce6f8378230ca WatchSource:0}: Error finding container 224863ed2625caf14a09031edc6fc0dda69fdeb66afe26f4d73ce6f8378230ca: Status 404 returned error can't find the container with id 224863ed2625caf14a09031edc6fc0dda69fdeb66afe26f4d73ce6f8378230ca Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.140429 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc91ace6-aab5-4a5c-9549-131d3303f5d5","Type":"ContainerStarted","Data":"41e37164f2490f0547ded0139a6ba0102e6e32485b2c8d3e197390fcef8ddd6a"} Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.147036 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerStarted","Data":"ebe0f20ab6e3ccdd587382d99ebb204edd65cbb0c28322d3a32fdf051aa4c586"} Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.147075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerStarted","Data":"224863ed2625caf14a09031edc6fc0dda69fdeb66afe26f4d73ce6f8378230ca"} Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.471527 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.810780 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7786888bd-6khpf" Oct 04 04:01:08 crc kubenswrapper[4726]: I1004 04:01:08.889089 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.160458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cc91ace6-aab5-4a5c-9549-131d3303f5d5","Type":"ContainerStarted","Data":"f9466a0a862330ec2936d698f089c8bb5d29a95446525a641b2852abce9478d4"} Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.161031 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.162800 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon-log" containerID="cri-o://3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6" gracePeriod=30 Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.163051 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerStarted","Data":"0b7ca876687086eddee9051dc937d773165f18b989eef425b82fe3936a73c3fd"} Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.163121 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" containerID="cri-o://267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4" gracePeriod=30 Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.180751 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.180733919 podStartE2EDuration="3.180733919s" podCreationTimestamp="2025-10-04 04:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:09.178929829 +0000 UTC m=+1243.353553032" watchObservedRunningTime="2025-10-04 04:01:09.180733919 +0000 UTC m=+1243.355357132" Oct 04 04:01:09 crc kubenswrapper[4726]: I1004 04:01:09.821524 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.189875 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerStarted","Data":"5907f9d22d48b07f89e20d72f4220bfd9eabd854e3a662e776eb675903f40eb4"} Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.189940 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.258073 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.258889 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="dnsmasq-dns" containerID="cri-o://e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107" gracePeriod=10 Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.435799 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.514278 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.535625 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-56bb546-6dhls" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.602473 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.602682 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57f84cf794-p9968" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api-log" containerID="cri-o://9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42" gracePeriod=30 Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.603053 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57f84cf794-p9968" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api" containerID="cri-o://26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55" gracePeriod=30 Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.810681 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.950487 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.950582 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.950653 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.950708 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.950746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjz57\" (UniqueName: \"kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.951557 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0\") pod \"2bc61ad6-c85f-45e3-9167-095574996c61\" (UID: \"2bc61ad6-c85f-45e3-9167-095574996c61\") " Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.959295 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57" (OuterVolumeSpecName: "kube-api-access-sjz57") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "kube-api-access-sjz57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.995153 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:10 crc kubenswrapper[4726]: I1004 04:01:10.997220 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.003904 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.006785 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.007711 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config" (OuterVolumeSpecName: "config") pod "2bc61ad6-c85f-45e3-9167-095574996c61" (UID: "2bc61ad6-c85f-45e3-9167-095574996c61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053666 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053700 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053711 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053723 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053734 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjz57\" (UniqueName: \"kubernetes.io/projected/2bc61ad6-c85f-45e3-9167-095574996c61-kube-api-access-sjz57\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.053744 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc61ad6-c85f-45e3-9167-095574996c61-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.200009 4726 generic.go:334] "Generic (PLEG): container finished" podID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerID="9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42" exitCode=143 Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.200853 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerDied","Data":"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42"} Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.203073 4726 generic.go:334] "Generic (PLEG): container finished" podID="2bc61ad6-c85f-45e3-9167-095574996c61" containerID="e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107" exitCode=0 Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.203718 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.203906 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="cinder-scheduler" containerID="cri-o://8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45" gracePeriod=30 Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.203908 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" event={"ID":"2bc61ad6-c85f-45e3-9167-095574996c61","Type":"ContainerDied","Data":"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107"} Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.204653 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-d72hr" event={"ID":"2bc61ad6-c85f-45e3-9167-095574996c61","Type":"ContainerDied","Data":"8a15b5a1e87576a121dd5db87b85b018355c4e0f513a8af3dfa231ab4b61049f"} Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.204690 4726 scope.go:117] "RemoveContainer" containerID="e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.204013 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="probe" containerID="cri-o://eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83" gracePeriod=30 Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.227165 4726 scope.go:117] "RemoveContainer" containerID="d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.243757 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.253480 4726 scope.go:117] "RemoveContainer" containerID="e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107" Oct 04 04:01:11 crc kubenswrapper[4726]: E1004 04:01:11.256738 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107\": container with ID starting with e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107 not found: ID does not exist" containerID="e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.256778 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107"} err="failed to get container status \"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107\": rpc error: code = NotFound desc = could not find container \"e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107\": container with ID starting with e2264a2563f9a7c8cd9bd9db6a1caa8aa2b6ad04a0a4eef78118021bc07fb107 not found: ID does not exist" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.256805 4726 scope.go:117] "RemoveContainer" containerID="d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8" Oct 04 04:01:11 crc kubenswrapper[4726]: E1004 04:01:11.257257 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8\": container with ID starting with d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8 not found: ID does not exist" containerID="d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.257279 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8"} err="failed to get container status \"d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8\": rpc error: code = NotFound desc = could not find container \"d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8\": container with ID starting with d50f93ec4d3ead5c59987b5314fe83a0a836fc4a279307a392214232515318f8 not found: ID does not exist" Oct 04 04:01:11 crc kubenswrapper[4726]: I1004 04:01:11.262826 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-d72hr"] Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.213650 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerStarted","Data":"c2d3f3ad2f8c4fd201a289d1a457534c5d5e14558055932155dcdf62b659b9ce"} Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.215175 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.217567 4726 generic.go:334] "Generic (PLEG): container finished" podID="626168ca-767c-43ae-8836-b2c75a8a7286" containerID="eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83" exitCode=0 Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.217620 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerDied","Data":"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83"} Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.239623 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.515447449 podStartE2EDuration="6.239592111s" podCreationTimestamp="2025-10-04 04:01:06 +0000 UTC" firstStartedPulling="2025-10-04 04:01:07.364328548 +0000 UTC m=+1241.538951771" lastFinishedPulling="2025-10-04 04:01:11.08847323 +0000 UTC m=+1245.263096433" observedRunningTime="2025-10-04 04:01:12.234231491 +0000 UTC m=+1246.408854754" watchObservedRunningTime="2025-10-04 04:01:12.239592111 +0000 UTC m=+1246.414215364" Oct 04 04:01:12 crc kubenswrapper[4726]: I1004 04:01:12.513968 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" path="/var/lib/kubelet/pods/2bc61ad6-c85f-45e3-9167-095574996c61/volumes" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.100292 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196443 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196495 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196545 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl4ck\" (UniqueName: \"kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196693 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196777 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.196838 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom\") pod \"626168ca-767c-43ae-8836-b2c75a8a7286\" (UID: \"626168ca-767c-43ae-8836-b2c75a8a7286\") " Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.197321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.203697 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck" (OuterVolumeSpecName: "kube-api-access-zl4ck") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "kube-api-access-zl4ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.204277 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts" (OuterVolumeSpecName: "scripts") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.205228 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.241857 4726 generic.go:334] "Generic (PLEG): container finished" podID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerID="267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4" exitCode=0 Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.241935 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerDied","Data":"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4"} Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.243807 4726 generic.go:334] "Generic (PLEG): container finished" podID="626168ca-767c-43ae-8836-b2c75a8a7286" containerID="8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45" exitCode=0 Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.243870 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerDied","Data":"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45"} Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.243896 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"626168ca-767c-43ae-8836-b2c75a8a7286","Type":"ContainerDied","Data":"cbdae37efbee918a8fd065050bfaa7fa8abc27268187a638abd0b0370aa3d9b7"} Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.243900 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.243912 4726 scope.go:117] "RemoveContainer" containerID="eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.266245 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.294544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data" (OuterVolumeSpecName: "config-data") pod "626168ca-767c-43ae-8836-b2c75a8a7286" (UID: "626168ca-767c-43ae-8836-b2c75a8a7286"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302137 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302162 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302173 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl4ck\" (UniqueName: \"kubernetes.io/projected/626168ca-767c-43ae-8836-b2c75a8a7286-kube-api-access-zl4ck\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302181 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302191 4726 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/626168ca-767c-43ae-8836-b2c75a8a7286-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.302199 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/626168ca-767c-43ae-8836-b2c75a8a7286-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.321974 4726 scope.go:117] "RemoveContainer" containerID="8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.356430 4726 scope.go:117] "RemoveContainer" containerID="eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.357904 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83\": container with ID starting with eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83 not found: ID does not exist" containerID="eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.357956 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83"} err="failed to get container status \"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83\": rpc error: code = NotFound desc = could not find container \"eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83\": container with ID starting with eaa75305e9aab5dd7ef448402067f0b99e46fa86d339c967bd9215c47b0dcf83 not found: ID does not exist" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.357982 4726 scope.go:117] "RemoveContainer" containerID="8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.358522 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45\": container with ID starting with 8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45 not found: ID does not exist" containerID="8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.358581 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45"} err="failed to get container status \"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45\": rpc error: code = NotFound desc = could not find container \"8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45\": container with ID starting with 8ece00f399da5be9684acb0b4576aaa558f1c3e4f1acd957e14adca11d5e2d45 not found: ID does not exist" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.581037 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.588376 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613143 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.613533 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="dnsmasq-dns" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613550 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="dnsmasq-dns" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.613574 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="probe" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613581 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="probe" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.613590 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="init" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613596 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="init" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.613609 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="cinder-scheduler" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613616 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="cinder-scheduler" Oct 04 04:01:13 crc kubenswrapper[4726]: E1004 04:01:13.613646 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db67cb41-86ca-41f3-b012-1bfc33e84a37" containerName="keystone-cron" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613654 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db67cb41-86ca-41f3-b012-1bfc33e84a37" containerName="keystone-cron" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613824 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="probe" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613843 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db67cb41-86ca-41f3-b012-1bfc33e84a37" containerName="keystone-cron" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613860 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc61ad6-c85f-45e3-9167-095574996c61" containerName="dnsmasq-dns" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.613868 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" containerName="cinder-scheduler" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.614808 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.617672 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.626536 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.708707 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fffz\" (UniqueName: \"kubernetes.io/projected/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-kube-api-access-5fffz\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.708838 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.708864 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.708888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.708947 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.709002 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.734695 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.792432 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836550 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836597 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836627 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fffz\" (UniqueName: \"kubernetes.io/projected/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-kube-api-access-5fffz\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836759 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.836782 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.841658 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.843740 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.845554 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.856936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-scripts\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.860806 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-config-data\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.869600 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fffz\" (UniqueName: \"kubernetes.io/projected/2105d235-5b6d-40ea-80d4-0f2033e0bf5d-kube-api-access-5fffz\") pod \"cinder-scheduler-0\" (UID: \"2105d235-5b6d-40ea-80d4-0f2033e0bf5d\") " pod="openstack/cinder-scheduler-0" Oct 04 04:01:13 crc kubenswrapper[4726]: I1004 04:01:13.946021 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.213033 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.262136 4726 generic.go:334] "Generic (PLEG): container finished" podID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerID="26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55" exitCode=0 Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.262219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerDied","Data":"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55"} Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.262250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f84cf794-p9968" event={"ID":"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c","Type":"ContainerDied","Data":"1b8ef4a2661a0d81c8484d9532e4bdc6f7900f61342750a536febba58fe42933"} Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.262272 4726 scope.go:117] "RemoveContainer" containerID="26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.262379 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f84cf794-p9968" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.295888 4726 scope.go:117] "RemoveContainer" containerID="9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.314846 4726 scope.go:117] "RemoveContainer" containerID="26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55" Oct 04 04:01:14 crc kubenswrapper[4726]: E1004 04:01:14.315289 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55\": container with ID starting with 26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55 not found: ID does not exist" containerID="26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.315345 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55"} err="failed to get container status \"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55\": rpc error: code = NotFound desc = could not find container \"26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55\": container with ID starting with 26e0477c31d07d9477d143b263fb7a2db5eae5978ef3c29f29bd54deff7ebf55 not found: ID does not exist" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.315377 4726 scope.go:117] "RemoveContainer" containerID="9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42" Oct 04 04:01:14 crc kubenswrapper[4726]: E1004 04:01:14.315700 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42\": container with ID starting with 9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42 not found: ID does not exist" containerID="9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.315732 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42"} err="failed to get container status \"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42\": rpc error: code = NotFound desc = could not find container \"9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42\": container with ID starting with 9787fa19ebcdf9287881b453f3fdde59c77cb93bdf3ad03798acf9720c4dde42 not found: ID does not exist" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.349461 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom\") pod \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.349522 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data\") pod \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.349572 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs\") pod \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.349596 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffvs2\" (UniqueName: \"kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2\") pod \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.349643 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle\") pod \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\" (UID: \"3f9c24f0-920b-4a44-9370-dbbdbe9ac69c\") " Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.350200 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs" (OuterVolumeSpecName: "logs") pod "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" (UID: "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.354005 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" (UID: "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.354443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2" (OuterVolumeSpecName: "kube-api-access-ffvs2") pod "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" (UID: "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c"). InnerVolumeSpecName "kube-api-access-ffvs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.383831 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" (UID: "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.419524 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data" (OuterVolumeSpecName: "config-data") pod "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" (UID: "3f9c24f0-920b-4a44-9370-dbbdbe9ac69c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.436135 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.451562 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.451675 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffvs2\" (UniqueName: \"kubernetes.io/projected/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-kube-api-access-ffvs2\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.451799 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.451857 4726 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.451909 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.513419 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="626168ca-767c-43ae-8836-b2c75a8a7286" path="/var/lib/kubelet/pods/626168ca-767c-43ae-8836-b2c75a8a7286/volumes" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.515650 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-79b54b49b-mx2kb" Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.598542 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:01:14 crc kubenswrapper[4726]: I1004 04:01:14.616577 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57f84cf794-p9968"] Oct 04 04:01:15 crc kubenswrapper[4726]: I1004 04:01:15.282564 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2105d235-5b6d-40ea-80d4-0f2033e0bf5d","Type":"ContainerStarted","Data":"8e2a40febc958acab700bf8251f2b9a1a7170774b7875e95b8dad922029780df"} Oct 04 04:01:15 crc kubenswrapper[4726]: I1004 04:01:15.282607 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2105d235-5b6d-40ea-80d4-0f2033e0bf5d","Type":"ContainerStarted","Data":"d7ef4dec2c2742a74a7993dd6740d3a20f4f01d26efd0ddcd18b64be252beec5"} Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.018032 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5bd64b8469-btvtm" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.118680 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.118919 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd6b788cb-dd5g5" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-api" containerID="cri-o://78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604" gracePeriod=30 Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.119462 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd6b788cb-dd5g5" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-httpd" containerID="cri-o://911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f" gracePeriod=30 Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.303663 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:01:16 crc kubenswrapper[4726]: E1004 04:01:16.304529 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.304544 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api" Oct 04 04:01:16 crc kubenswrapper[4726]: E1004 04:01:16.304562 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api-log" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.304568 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api-log" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.304839 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api-log" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.304855 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" containerName="barbican-api" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.305716 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.307550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2105d235-5b6d-40ea-80d4-0f2033e0bf5d","Type":"ContainerStarted","Data":"1830f43e9abe6775ddb5780bd4b267e551e13e33ca621e56c4fc3c1130ce2041"} Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.308660 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.308846 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.309459 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-4knrf" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.328789 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.369059 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.369041974 podStartE2EDuration="3.369041974s" podCreationTimestamp="2025-10-04 04:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:16.367365228 +0000 UTC m=+1250.541988441" watchObservedRunningTime="2025-10-04 04:01:16.369041974 +0000 UTC m=+1250.543665187" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.430400 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config-secret\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.430529 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkxld\" (UniqueName: \"kubernetes.io/projected/f592dceb-e227-48ea-a6d6-0b9513ec5b46-kube-api-access-tkxld\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.430633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.430649 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.513697 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f9c24f0-920b-4a44-9370-dbbdbe9ac69c" path="/var/lib/kubelet/pods/3f9c24f0-920b-4a44-9370-dbbdbe9ac69c/volumes" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.532741 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.532792 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.532823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config-secret\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.532984 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkxld\" (UniqueName: \"kubernetes.io/projected/f592dceb-e227-48ea-a6d6-0b9513ec5b46-kube-api-access-tkxld\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.533834 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.538863 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.555622 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f592dceb-e227-48ea-a6d6-0b9513ec5b46-openstack-config-secret\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.560801 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkxld\" (UniqueName: \"kubernetes.io/projected/f592dceb-e227-48ea-a6d6-0b9513ec5b46-kube-api-access-tkxld\") pod \"openstackclient\" (UID: \"f592dceb-e227-48ea-a6d6-0b9513ec5b46\") " pod="openstack/openstackclient" Oct 04 04:01:16 crc kubenswrapper[4726]: I1004 04:01:16.688810 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:01:17 crc kubenswrapper[4726]: I1004 04:01:17.190660 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:01:17 crc kubenswrapper[4726]: W1004 04:01:17.194704 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf592dceb_e227_48ea_a6d6_0b9513ec5b46.slice/crio-b3c43881c843a4c3dcfa4af3bfa0274be1ec37793db455a6000f1165d944c668 WatchSource:0}: Error finding container b3c43881c843a4c3dcfa4af3bfa0274be1ec37793db455a6000f1165d944c668: Status 404 returned error can't find the container with id b3c43881c843a4c3dcfa4af3bfa0274be1ec37793db455a6000f1165d944c668 Oct 04 04:01:17 crc kubenswrapper[4726]: I1004 04:01:17.333419 4726 generic.go:334] "Generic (PLEG): container finished" podID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerID="911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f" exitCode=0 Oct 04 04:01:17 crc kubenswrapper[4726]: I1004 04:01:17.333518 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerDied","Data":"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f"} Oct 04 04:01:17 crc kubenswrapper[4726]: I1004 04:01:17.336888 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f592dceb-e227-48ea-a6d6-0b9513ec5b46","Type":"ContainerStarted","Data":"b3c43881c843a4c3dcfa4af3bfa0274be1ec37793db455a6000f1165d944c668"} Oct 04 04:01:18 crc kubenswrapper[4726]: I1004 04:01:18.526891 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:01:18 crc kubenswrapper[4726]: I1004 04:01:18.946963 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.226458 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5bdb856b6c-f5xfb"] Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.229135 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.232142 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.232282 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.232353 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.247086 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bdb856b6c-f5xfb"] Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298487 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-internal-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298552 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-log-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298629 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-run-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298671 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxks2\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-kube-api-access-hxks2\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298788 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-config-data\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298827 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-etc-swift\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298866 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-combined-ca-bundle\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.298907 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-public-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400038 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-public-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400151 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-internal-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400184 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-log-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400237 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-run-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400263 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxks2\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-kube-api-access-hxks2\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400664 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-run-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.400817 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/503e5303-a481-4184-9bb0-2369123d5267-log-httpd\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.401087 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-config-data\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.401192 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-etc-swift\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.401221 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-combined-ca-bundle\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.407362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-config-data\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.407408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-public-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.408973 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-combined-ca-bundle\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.424017 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/503e5303-a481-4184-9bb0-2369123d5267-internal-tls-certs\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.428264 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-etc-swift\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.432011 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxks2\" (UniqueName: \"kubernetes.io/projected/503e5303-a481-4184-9bb0-2369123d5267-kube-api-access-hxks2\") pod \"swift-proxy-5bdb856b6c-f5xfb\" (UID: \"503e5303-a481-4184-9bb0-2369123d5267\") " pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.550175 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.572044 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.572954 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-central-agent" containerID="cri-o://ebe0f20ab6e3ccdd587382d99ebb204edd65cbb0c28322d3a32fdf051aa4c586" gracePeriod=30 Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.572992 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-notification-agent" containerID="cri-o://0b7ca876687086eddee9051dc937d773165f18b989eef425b82fe3936a73c3fd" gracePeriod=30 Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.573005 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="proxy-httpd" containerID="cri-o://c2d3f3ad2f8c4fd201a289d1a457534c5d5e14558055932155dcdf62b659b9ce" gracePeriod=30 Oct 04 04:01:20 crc kubenswrapper[4726]: I1004 04:01:20.573013 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="sg-core" containerID="cri-o://5907f9d22d48b07f89e20d72f4220bfd9eabd854e3a662e776eb675903f40eb4" gracePeriod=30 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.253328 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bdb856b6c-f5xfb"] Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.291448 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.324076 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config\") pod \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.324176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jxh7\" (UniqueName: \"kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7\") pod \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.324209 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle\") pod \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.324248 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs\") pod \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.324300 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config\") pod \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\" (UID: \"d48a45ab-be06-4a10-b14c-ca3354afd6ff\") " Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.329756 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7" (OuterVolumeSpecName: "kube-api-access-5jxh7") pod "d48a45ab-be06-4a10-b14c-ca3354afd6ff" (UID: "d48a45ab-be06-4a10-b14c-ca3354afd6ff"). InnerVolumeSpecName "kube-api-access-5jxh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.343025 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d48a45ab-be06-4a10-b14c-ca3354afd6ff" (UID: "d48a45ab-be06-4a10-b14c-ca3354afd6ff"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.397081 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.397450 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-log" containerID="cri-o://617f7095f19e49a0c5a39d938c529b27e0fb840bb9bce6a684f4ee9eac81ea5b" gracePeriod=30 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.397902 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-httpd" containerID="cri-o://38b5892dcd0ab4e81a7fa3d1519031c715a6f637603d10a642538e3164101c96" gracePeriod=30 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.404232 4726 generic.go:334] "Generic (PLEG): container finished" podID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerID="78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604" exitCode=0 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.404307 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd6b788cb-dd5g5" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.404438 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerDied","Data":"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.404471 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd6b788cb-dd5g5" event={"ID":"d48a45ab-be06-4a10-b14c-ca3354afd6ff","Type":"ContainerDied","Data":"d02f490eed6321c0045320e6144dcfa0884cd5fbb1d6bede490fd17e184af498"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.404491 4726 scope.go:117] "RemoveContainer" containerID="911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.416384 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config" (OuterVolumeSpecName: "config") pod "d48a45ab-be06-4a10-b14c-ca3354afd6ff" (UID: "d48a45ab-be06-4a10-b14c-ca3354afd6ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419570 4726 generic.go:334] "Generic (PLEG): container finished" podID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerID="c2d3f3ad2f8c4fd201a289d1a457534c5d5e14558055932155dcdf62b659b9ce" exitCode=0 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419617 4726 generic.go:334] "Generic (PLEG): container finished" podID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerID="5907f9d22d48b07f89e20d72f4220bfd9eabd854e3a662e776eb675903f40eb4" exitCode=2 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419632 4726 generic.go:334] "Generic (PLEG): container finished" podID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerID="ebe0f20ab6e3ccdd587382d99ebb204edd65cbb0c28322d3a32fdf051aa4c586" exitCode=0 Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerDied","Data":"c2d3f3ad2f8c4fd201a289d1a457534c5d5e14558055932155dcdf62b659b9ce"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419750 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerDied","Data":"5907f9d22d48b07f89e20d72f4220bfd9eabd854e3a662e776eb675903f40eb4"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.419787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerDied","Data":"ebe0f20ab6e3ccdd587382d99ebb204edd65cbb0c28322d3a32fdf051aa4c586"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.423931 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" event={"ID":"503e5303-a481-4184-9bb0-2369123d5267","Type":"ContainerStarted","Data":"7c3944b623f865e7e1fea6c2335a5d9f2c8383d60d222aae2c28082901ab76b4"} Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.427079 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.427130 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.427143 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jxh7\" (UniqueName: \"kubernetes.io/projected/d48a45ab-be06-4a10-b14c-ca3354afd6ff-kube-api-access-5jxh7\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.445801 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d48a45ab-be06-4a10-b14c-ca3354afd6ff" (UID: "d48a45ab-be06-4a10-b14c-ca3354afd6ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.471634 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d48a45ab-be06-4a10-b14c-ca3354afd6ff" (UID: "d48a45ab-be06-4a10-b14c-ca3354afd6ff"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.528748 4726 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.528985 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d48a45ab-be06-4a10-b14c-ca3354afd6ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.615431 4726 scope.go:117] "RemoveContainer" containerID="78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.638697 4726 scope.go:117] "RemoveContainer" containerID="911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f" Oct 04 04:01:21 crc kubenswrapper[4726]: E1004 04:01:21.641184 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f\": container with ID starting with 911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f not found: ID does not exist" containerID="911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.641284 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f"} err="failed to get container status \"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f\": rpc error: code = NotFound desc = could not find container \"911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f\": container with ID starting with 911ce73d17b85af0df848eeca7a2c7ef5ec8aedd10179d39692da68f6701807f not found: ID does not exist" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.641360 4726 scope.go:117] "RemoveContainer" containerID="78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604" Oct 04 04:01:21 crc kubenswrapper[4726]: E1004 04:01:21.642808 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604\": container with ID starting with 78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604 not found: ID does not exist" containerID="78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.642843 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604"} err="failed to get container status \"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604\": rpc error: code = NotFound desc = could not find container \"78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604\": container with ID starting with 78226ab28d416c9d1416342d290e11b824c6cef5e4aac47d2b6eb601df29f604 not found: ID does not exist" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.674328 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.717563 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f697b85fb-mcpgk" Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.768550 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:01:21 crc kubenswrapper[4726]: I1004 04:01:21.777891 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5dd6b788cb-dd5g5"] Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.438876 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" event={"ID":"503e5303-a481-4184-9bb0-2369123d5267","Type":"ContainerStarted","Data":"948849b06e297710f6bda28b20f13987c585969b2caab76596ae1951d2893665"} Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.439934 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" event={"ID":"503e5303-a481-4184-9bb0-2369123d5267","Type":"ContainerStarted","Data":"9e5644e6bcb0354d51e4d363f321359e8b54e607f8bd299a4893e7d40500e1be"} Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.439965 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.444493 4726 generic.go:334] "Generic (PLEG): container finished" podID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerID="617f7095f19e49a0c5a39d938c529b27e0fb840bb9bce6a684f4ee9eac81ea5b" exitCode=143 Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.444611 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerDied","Data":"617f7095f19e49a0c5a39d938c529b27e0fb840bb9bce6a684f4ee9eac81ea5b"} Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.464584 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" podStartSLOduration=2.46457023 podStartE2EDuration="2.46457023s" podCreationTimestamp="2025-10-04 04:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:22.45847646 +0000 UTC m=+1256.633099673" watchObservedRunningTime="2025-10-04 04:01:22.46457023 +0000 UTC m=+1256.639193433" Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.513000 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" path="/var/lib/kubelet/pods/d48a45ab-be06-4a10-b14c-ca3354afd6ff/volumes" Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.778932 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.779550 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-httpd" containerID="cri-o://c25f53e7f0cfa68844c99391f0d8fcd32ad1bf0f2e03f9af9e411c28e91932d6" gracePeriod=30 Oct 04 04:01:22 crc kubenswrapper[4726]: I1004 04:01:22.781502 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-log" containerID="cri-o://62218961f079497875f721f630e4e5747abbeb1de19e905266446070263c456b" gracePeriod=30 Oct 04 04:01:23 crc kubenswrapper[4726]: I1004 04:01:23.464486 4726 generic.go:334] "Generic (PLEG): container finished" podID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerID="62218961f079497875f721f630e4e5747abbeb1de19e905266446070263c456b" exitCode=143 Oct 04 04:01:23 crc kubenswrapper[4726]: I1004 04:01:23.464551 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerDied","Data":"62218961f079497875f721f630e4e5747abbeb1de19e905266446070263c456b"} Oct 04 04:01:23 crc kubenswrapper[4726]: I1004 04:01:23.465273 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:23 crc kubenswrapper[4726]: I1004 04:01:23.735227 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:01:24 crc kubenswrapper[4726]: I1004 04:01:24.280754 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:01:25 crc kubenswrapper[4726]: I1004 04:01:25.495986 4726 generic.go:334] "Generic (PLEG): container finished" podID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerID="0b7ca876687086eddee9051dc937d773165f18b989eef425b82fe3936a73c3fd" exitCode=0 Oct 04 04:01:25 crc kubenswrapper[4726]: I1004 04:01:25.496099 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerDied","Data":"0b7ca876687086eddee9051dc937d773165f18b989eef425b82fe3936a73c3fd"} Oct 04 04:01:25 crc kubenswrapper[4726]: I1004 04:01:25.501352 4726 generic.go:334] "Generic (PLEG): container finished" podID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerID="38b5892dcd0ab4e81a7fa3d1519031c715a6f637603d10a642538e3164101c96" exitCode=0 Oct 04 04:01:25 crc kubenswrapper[4726]: I1004 04:01:25.501565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerDied","Data":"38b5892dcd0ab4e81a7fa3d1519031c715a6f637603d10a642538e3164101c96"} Oct 04 04:01:26 crc kubenswrapper[4726]: I1004 04:01:26.512036 4726 generic.go:334] "Generic (PLEG): container finished" podID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerID="c25f53e7f0cfa68844c99391f0d8fcd32ad1bf0f2e03f9af9e411c28e91932d6" exitCode=0 Oct 04 04:01:26 crc kubenswrapper[4726]: I1004 04:01:26.515216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerDied","Data":"c25f53e7f0cfa68844c99391f0d8fcd32ad1bf0f2e03f9af9e411c28e91932d6"} Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.273858 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.387341 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469608 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469715 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vc9k8\" (UniqueName: \"kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469820 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469857 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469879 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.469928 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd\") pod \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\" (UID: \"71eeaf87-9cec-4ba0-9702-99bacffc46a3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.470048 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.470485 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.470740 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.470838 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.470931 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.471628 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.471647 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71eeaf87-9cec-4ba0-9702-99bacffc46a3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.476636 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs" (OuterVolumeSpecName: "logs") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.476818 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8" (OuterVolumeSpecName: "kube-api-access-vc9k8") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "kube-api-access-vc9k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.479995 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts" (OuterVolumeSpecName: "scripts") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.479992 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts" (OuterVolumeSpecName: "scripts") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.507341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.529689 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f592dceb-e227-48ea-a6d6-0b9513ec5b46","Type":"ContainerStarted","Data":"802e17948852826902e846042867d746a3faa8f5f16e0bcd9baf3d4957a8faf8"} Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.533040 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d805b34d-4c76-4e56-96d9-c19c6b3320b3","Type":"ContainerDied","Data":"a0d842fa13a2722bb75353ca431906776ad5260502a7c1254f23c450e66fcaf4"} Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.533056 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.533151 4726 scope.go:117] "RemoveContainer" containerID="38b5892dcd0ab4e81a7fa3d1519031c715a6f637603d10a642538e3164101c96" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.538252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71eeaf87-9cec-4ba0-9702-99bacffc46a3","Type":"ContainerDied","Data":"224863ed2625caf14a09031edc6fc0dda69fdeb66afe26f4d73ce6f8378230ca"} Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.538297 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.544411 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.545342 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.708206404 podStartE2EDuration="12.545328274s" podCreationTimestamp="2025-10-04 04:01:16 +0000 UTC" firstStartedPulling="2025-10-04 04:01:17.196952987 +0000 UTC m=+1251.371576200" lastFinishedPulling="2025-10-04 04:01:28.034074857 +0000 UTC m=+1262.208698070" observedRunningTime="2025-10-04 04:01:28.544634805 +0000 UTC m=+1262.719258018" watchObservedRunningTime="2025-10-04 04:01:28.545328274 +0000 UTC m=+1262.719951487" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.571441 4726 scope.go:117] "RemoveContainer" containerID="617f7095f19e49a0c5a39d938c529b27e0fb840bb9bce6a684f4ee9eac81ea5b" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.572635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.572697 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.572721 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dczkr\" (UniqueName: \"kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.572764 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.573086 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.574037 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data\") pod \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\" (UID: \"d805b34d-4c76-4e56-96d9-c19c6b3320b3\") " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575329 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575349 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575361 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575372 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vc9k8\" (UniqueName: \"kubernetes.io/projected/71eeaf87-9cec-4ba0-9702-99bacffc46a3-kube-api-access-vc9k8\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575384 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575395 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d805b34d-4c76-4e56-96d9-c19c6b3320b3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.575419 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.576046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.576857 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr" (OuterVolumeSpecName: "kube-api-access-dczkr") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "kube-api-access-dczkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.600810 4726 scope.go:117] "RemoveContainer" containerID="c2d3f3ad2f8c4fd201a289d1a457534c5d5e14558055932155dcdf62b659b9ce" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.606564 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.608831 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.621228 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data" (OuterVolumeSpecName: "config-data") pod "71eeaf87-9cec-4ba0-9702-99bacffc46a3" (UID: "71eeaf87-9cec-4ba0-9702-99bacffc46a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.627310 4726 scope.go:117] "RemoveContainer" containerID="5907f9d22d48b07f89e20d72f4220bfd9eabd854e3a662e776eb675903f40eb4" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.635060 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data" (OuterVolumeSpecName: "config-data") pod "d805b34d-4c76-4e56-96d9-c19c6b3320b3" (UID: "d805b34d-4c76-4e56-96d9-c19c6b3320b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.648340 4726 scope.go:117] "RemoveContainer" containerID="0b7ca876687086eddee9051dc937d773165f18b989eef425b82fe3936a73c3fd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.669042 4726 scope.go:117] "RemoveContainer" containerID="ebe0f20ab6e3ccdd587382d99ebb204edd65cbb0c28322d3a32fdf051aa4c586" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677412 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677441 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677452 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dczkr\" (UniqueName: \"kubernetes.io/projected/d805b34d-4c76-4e56-96d9-c19c6b3320b3-kube-api-access-dczkr\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677466 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677475 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d805b34d-4c76-4e56-96d9-c19c6b3320b3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.677484 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71eeaf87-9cec-4ba0-9702-99bacffc46a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.698024 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.779401 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.871577 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.881209 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.897005 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.909537 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.918438 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.918914 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.918931 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.918963 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-central-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.918975 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-central-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.918993 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="sg-core" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919001 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="sg-core" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.919015 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-notification-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919024 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-notification-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.919041 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919050 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.919068 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-log" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919076 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-log" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.919089 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-api" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919097 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-api" Oct 04 04:01:28 crc kubenswrapper[4726]: E1004 04:01:28.919130 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="proxy-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919138 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="proxy-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919351 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-notification-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919372 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919387 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-api" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919398 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" containerName="glance-log" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919417 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="ceilometer-central-agent" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919427 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="proxy-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919450 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48a45ab-be06-4a10-b14c-ca3354afd6ff" containerName="neutron-httpd" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.919469 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" containerName="sg-core" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.920643 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.931181 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.933159 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.933905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.936516 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.936740 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.940551 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.941878 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.957255 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983645 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c852l\" (UniqueName: \"kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983749 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983781 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983812 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983866 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983902 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:28 crc kubenswrapper[4726]: I1004 04:01:28.983930 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.085952 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086008 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086051 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9tvk\" (UniqueName: \"kubernetes.io/projected/636ad703-b66b-42f6-aca7-00900d0fad9b-kube-api-access-p9tvk\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086091 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086136 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086169 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086198 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086280 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086317 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086354 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086392 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086414 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086438 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086464 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c852l\" (UniqueName: \"kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.086899 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.087568 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.091814 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.092475 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.096914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.097070 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.112669 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c852l\" (UniqueName: \"kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l\") pod \"ceilometer-0\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.187888 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188207 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188225 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188244 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188283 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188306 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188341 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9tvk\" (UniqueName: \"kubernetes.io/projected/636ad703-b66b-42f6-aca7-00900d0fad9b-kube-api-access-p9tvk\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.188383 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.189054 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.191716 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.191813 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.192167 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/636ad703-b66b-42f6-aca7-00900d0fad9b-logs\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.196144 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.203611 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.203725 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636ad703-b66b-42f6-aca7-00900d0fad9b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.213859 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9tvk\" (UniqueName: \"kubernetes.io/projected/636ad703-b66b-42f6-aca7-00900d0fad9b-kube-api-access-p9tvk\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.228537 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"636ad703-b66b-42f6-aca7-00900d0fad9b\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.290265 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.312095 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.328272 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391078 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjm9s\" (UniqueName: \"kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391157 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391196 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391216 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391276 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391317 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391383 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.391408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run\") pod \"adb80860-f5c3-417b-a55f-cc94f41568e3\" (UID: \"adb80860-f5c3-417b-a55f-cc94f41568e3\") " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.394620 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.395852 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs" (OuterVolumeSpecName: "logs") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.401726 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts" (OuterVolumeSpecName: "scripts") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.409066 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s" (OuterVolumeSpecName: "kube-api-access-mjm9s") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "kube-api-access-mjm9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.414260 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.462602 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498035 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498061 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498070 4726 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/adb80860-f5c3-417b-a55f-cc94f41568e3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498079 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjm9s\" (UniqueName: \"kubernetes.io/projected/adb80860-f5c3-417b-a55f-cc94f41568e3-kube-api-access-mjm9s\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498090 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.498132 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.511402 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data" (OuterVolumeSpecName: "config-data") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.516910 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "adb80860-f5c3-417b-a55f-cc94f41568e3" (UID: "adb80860-f5c3-417b-a55f-cc94f41568e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.528489 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.554546 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.554541 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"adb80860-f5c3-417b-a55f-cc94f41568e3","Type":"ContainerDied","Data":"41ebf9a89568f47326f5d12e1d7ce2d38b4283103aa570ca55f58c3c3c510afe"} Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.554719 4726 scope.go:117] "RemoveContainer" containerID="c25f53e7f0cfa68844c99391f0d8fcd32ad1bf0f2e03f9af9e411c28e91932d6" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.597631 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.599061 4726 scope.go:117] "RemoveContainer" containerID="62218961f079497875f721f630e4e5747abbeb1de19e905266446070263c456b" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.601043 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.601069 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adb80860-f5c3-417b-a55f-cc94f41568e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.601079 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.604702 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.626788 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:29 crc kubenswrapper[4726]: E1004 04:01:29.627178 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-httpd" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.627191 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-httpd" Oct 04 04:01:29 crc kubenswrapper[4726]: E1004 04:01:29.627203 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-log" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.627209 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-log" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.627404 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-httpd" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.627421 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" containerName="glance-log" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.628325 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.630784 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.631045 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.642598 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804642 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-scripts\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804690 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-config-data\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804723 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804772 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-logs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804791 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804868 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfcw2\" (UniqueName: \"kubernetes.io/projected/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-kube-api-access-rfcw2\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.804913 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.865888 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:29 crc kubenswrapper[4726]: W1004 04:01:29.869624 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85cb2d95_8133_4c7f_8ee4_ca4683e39160.slice/crio-79e55aba2de63ee9fa447938a34a9a5513445b560c5c949ef4ec456c65d3d5d3 WatchSource:0}: Error finding container 79e55aba2de63ee9fa447938a34a9a5513445b560c5c949ef4ec456c65d3d5d3: Status 404 returned error can't find the container with id 79e55aba2de63ee9fa447938a34a9a5513445b560c5c949ef4ec456c65d3d5d3 Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908427 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-logs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908478 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfcw2\" (UniqueName: \"kubernetes.io/projected/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-kube-api-access-rfcw2\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908613 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908638 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-scripts\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908662 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-config-data\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.908901 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.909072 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.909445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-logs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.909811 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.916881 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-config-data\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.917481 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.917833 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.917885 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-scripts\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.930628 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfcw2\" (UniqueName: \"kubernetes.io/projected/bae4a133-24a9-4ba4-8050-5e1a15bc8ed3-kube-api-access-rfcw2\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.944050 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3\") " pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.958257 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:01:29 crc kubenswrapper[4726]: I1004 04:01:29.975986 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.517029 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71eeaf87-9cec-4ba0-9702-99bacffc46a3" path="/var/lib/kubelet/pods/71eeaf87-9cec-4ba0-9702-99bacffc46a3/volumes" Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.518600 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adb80860-f5c3-417b-a55f-cc94f41568e3" path="/var/lib/kubelet/pods/adb80860-f5c3-417b-a55f-cc94f41568e3/volumes" Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.519933 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d805b34d-4c76-4e56-96d9-c19c6b3320b3" path="/var/lib/kubelet/pods/d805b34d-4c76-4e56-96d9-c19c6b3320b3/volumes" Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.538578 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.564615 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.566332 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.580035 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3","Type":"ContainerStarted","Data":"e1b1b7e7d5f38a50c1ce2cda388e00d46cca118e9ff7e9d946e0a4f7c090d237"} Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.582048 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"636ad703-b66b-42f6-aca7-00900d0fad9b","Type":"ContainerStarted","Data":"c2fcaf1640ab1d27762c246619d9eb913f47c709bf5b14c2d059c5c51a353525"} Oct 04 04:01:30 crc kubenswrapper[4726]: I1004 04:01:30.582915 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerStarted","Data":"79e55aba2de63ee9fa447938a34a9a5513445b560c5c949ef4ec456c65d3d5d3"} Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.355093 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.600018 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3","Type":"ContainerStarted","Data":"2b7fccf6be1117df2bac2bbfb0eb7f9d1822d3fd54a4536a342cd5b5c469c41b"} Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.602236 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"636ad703-b66b-42f6-aca7-00900d0fad9b","Type":"ContainerStarted","Data":"449fd60dc5a80f7690f3e8b0bf2fe75ae0ad0ee007bb9ae31894574303cfbcb5"} Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.602285 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"636ad703-b66b-42f6-aca7-00900d0fad9b","Type":"ContainerStarted","Data":"4505d9acebfc7977b55dabcec4f17f621bea83ed4aac699789cd35ac3cdf6bde"} Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.606527 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerStarted","Data":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} Oct 04 04:01:31 crc kubenswrapper[4726]: I1004 04:01:31.627203 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.627178485 podStartE2EDuration="3.627178485s" podCreationTimestamp="2025-10-04 04:01:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:31.622791993 +0000 UTC m=+1265.797415206" watchObservedRunningTime="2025-10-04 04:01:31.627178485 +0000 UTC m=+1265.801801698" Oct 04 04:01:32 crc kubenswrapper[4726]: I1004 04:01:32.617121 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bae4a133-24a9-4ba4-8050-5e1a15bc8ed3","Type":"ContainerStarted","Data":"ee661ffc483c3fed60d2cc2f876bbefeaecd8f941fe9d8ba51850e5eb777f24b"} Oct 04 04:01:32 crc kubenswrapper[4726]: I1004 04:01:32.618840 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerStarted","Data":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} Oct 04 04:01:32 crc kubenswrapper[4726]: I1004 04:01:32.644461 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.644436573 podStartE2EDuration="3.644436573s" podCreationTimestamp="2025-10-04 04:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:01:32.636479202 +0000 UTC m=+1266.811102445" watchObservedRunningTime="2025-10-04 04:01:32.644436573 +0000 UTC m=+1266.819059786" Oct 04 04:01:33 crc kubenswrapper[4726]: I1004 04:01:33.629963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerStarted","Data":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} Oct 04 04:01:33 crc kubenswrapper[4726]: I1004 04:01:33.735076 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b6cffb848-k9fxk" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:01:33 crc kubenswrapper[4726]: I1004 04:01:33.735236 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.666888 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerStarted","Data":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.667408 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.667192 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="proxy-httpd" containerID="cri-o://40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" gracePeriod=30 Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.667236 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-notification-agent" containerID="cri-o://286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" gracePeriod=30 Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.667257 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="sg-core" containerID="cri-o://ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" gracePeriod=30 Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.667070 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-central-agent" containerID="cri-o://afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" gracePeriod=30 Oct 04 04:01:35 crc kubenswrapper[4726]: I1004 04:01:35.702606 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.898919824 podStartE2EDuration="7.702580404s" podCreationTimestamp="2025-10-04 04:01:28 +0000 UTC" firstStartedPulling="2025-10-04 04:01:29.871923275 +0000 UTC m=+1264.046546488" lastFinishedPulling="2025-10-04 04:01:34.675583845 +0000 UTC m=+1268.850207068" observedRunningTime="2025-10-04 04:01:35.693687437 +0000 UTC m=+1269.868310670" watchObservedRunningTime="2025-10-04 04:01:35.702580404 +0000 UTC m=+1269.877203617" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.420708 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.537337 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.537438 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.537608 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.538289 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c852l\" (UniqueName: \"kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.538366 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.538418 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.538485 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd\") pod \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\" (UID: \"85cb2d95-8133-4c7f-8ee4-ca4683e39160\") " Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.539078 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.539117 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.539241 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.539256 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85cb2d95-8133-4c7f-8ee4-ca4683e39160-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.543136 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts" (OuterVolumeSpecName: "scripts") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.545852 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l" (OuterVolumeSpecName: "kube-api-access-c852l") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "kube-api-access-c852l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.572024 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.632554 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.640341 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c852l\" (UniqueName: \"kubernetes.io/projected/85cb2d95-8133-4c7f-8ee4-ca4683e39160-kube-api-access-c852l\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.640373 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.640383 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.640391 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.644651 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data" (OuterVolumeSpecName: "config-data") pod "85cb2d95-8133-4c7f-8ee4-ca4683e39160" (UID: "85cb2d95-8133-4c7f-8ee4-ca4683e39160"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676012 4726 generic.go:334] "Generic (PLEG): container finished" podID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" exitCode=0 Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676050 4726 generic.go:334] "Generic (PLEG): container finished" podID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" exitCode=2 Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676065 4726 generic.go:334] "Generic (PLEG): container finished" podID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" exitCode=0 Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676078 4726 generic.go:334] "Generic (PLEG): container finished" podID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" exitCode=0 Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676067 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerDied","Data":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676213 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerDied","Data":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerDied","Data":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerDied","Data":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676264 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85cb2d95-8133-4c7f-8ee4-ca4683e39160","Type":"ContainerDied","Data":"79e55aba2de63ee9fa447938a34a9a5513445b560c5c949ef4ec456c65d3d5d3"} Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.676271 4726 scope.go:117] "RemoveContainer" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.703693 4726 scope.go:117] "RemoveContainer" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.707731 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.717531 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732218 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732267 4726 scope.go:117] "RemoveContainer" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.732650 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-notification-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732674 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-notification-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.732693 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="proxy-httpd" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732703 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="proxy-httpd" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.732735 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="sg-core" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732746 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="sg-core" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.732774 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-central-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.732784 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-central-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.733015 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-notification-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.733046 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="proxy-httpd" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.733069 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="sg-core" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.733083 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" containerName="ceilometer-central-agent" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.736374 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.738627 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.741411 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85cb2d95-8133-4c7f-8ee4-ca4683e39160-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.743383 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.748065 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.754438 4726 scope.go:117] "RemoveContainer" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842541 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842622 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842643 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842666 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842756 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.842774 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxqhb\" (UniqueName: \"kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.851549 4726 scope.go:117] "RemoveContainer" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.851923 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": container with ID starting with 40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741 not found: ID does not exist" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.851957 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} err="failed to get container status \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": rpc error: code = NotFound desc = could not find container \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": container with ID starting with 40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.851983 4726 scope.go:117] "RemoveContainer" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.852359 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": container with ID starting with ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12 not found: ID does not exist" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.852391 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} err="failed to get container status \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": rpc error: code = NotFound desc = could not find container \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": container with ID starting with ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.852408 4726 scope.go:117] "RemoveContainer" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.853112 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": container with ID starting with 286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a not found: ID does not exist" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853140 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} err="failed to get container status \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": rpc error: code = NotFound desc = could not find container \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": container with ID starting with 286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853158 4726 scope.go:117] "RemoveContainer" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: E1004 04:01:36.853471 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": container with ID starting with afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870 not found: ID does not exist" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853548 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} err="failed to get container status \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": rpc error: code = NotFound desc = could not find container \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": container with ID starting with afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853608 4726 scope.go:117] "RemoveContainer" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853848 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} err="failed to get container status \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": rpc error: code = NotFound desc = could not find container \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": container with ID starting with 40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.853867 4726 scope.go:117] "RemoveContainer" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854043 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} err="failed to get container status \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": rpc error: code = NotFound desc = could not find container \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": container with ID starting with ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854126 4726 scope.go:117] "RemoveContainer" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854367 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} err="failed to get container status \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": rpc error: code = NotFound desc = could not find container \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": container with ID starting with 286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854389 4726 scope.go:117] "RemoveContainer" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854576 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} err="failed to get container status \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": rpc error: code = NotFound desc = could not find container \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": container with ID starting with afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854661 4726 scope.go:117] "RemoveContainer" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.854984 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} err="failed to get container status \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": rpc error: code = NotFound desc = could not find container \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": container with ID starting with 40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855039 4726 scope.go:117] "RemoveContainer" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855327 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} err="failed to get container status \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": rpc error: code = NotFound desc = could not find container \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": container with ID starting with ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855401 4726 scope.go:117] "RemoveContainer" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855619 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} err="failed to get container status \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": rpc error: code = NotFound desc = could not find container \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": container with ID starting with 286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855639 4726 scope.go:117] "RemoveContainer" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855804 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} err="failed to get container status \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": rpc error: code = NotFound desc = could not find container \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": container with ID starting with afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.855879 4726 scope.go:117] "RemoveContainer" containerID="40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.856769 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741"} err="failed to get container status \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": rpc error: code = NotFound desc = could not find container \"40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741\": container with ID starting with 40967fcf35df1a0e94d3ba5b982b510df8aca06f44e6fb06e63361701a798741 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.856872 4726 scope.go:117] "RemoveContainer" containerID="ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.857184 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12"} err="failed to get container status \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": rpc error: code = NotFound desc = could not find container \"ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12\": container with ID starting with ca5290e689b2bdb7fd4aa6465cdb28a88e80c41dd7862c0b0a7733b7abbf4c12 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.857202 4726 scope.go:117] "RemoveContainer" containerID="286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.857523 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a"} err="failed to get container status \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": rpc error: code = NotFound desc = could not find container \"286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a\": container with ID starting with 286c7719f5fa9a8eaac37b30d8306e93b83c4457425d837e85fe3fddd84e2f1a not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.857545 4726 scope.go:117] "RemoveContainer" containerID="afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.857806 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870"} err="failed to get container status \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": rpc error: code = NotFound desc = could not find container \"afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870\": container with ID starting with afcff0fc684d3e8b63c99ded40ff3b1cd8e6f386ee710a081c7a6f1a70768870 not found: ID does not exist" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944658 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944714 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944732 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944751 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944821 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944840 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.944857 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxqhb\" (UniqueName: \"kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.945583 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.946785 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.950005 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.950087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.955408 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.955781 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:36 crc kubenswrapper[4726]: I1004 04:01:36.960835 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxqhb\" (UniqueName: \"kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb\") pod \"ceilometer-0\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " pod="openstack/ceilometer-0" Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.148731 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.647989 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:37 crc kubenswrapper[4726]: W1004 04:01:37.673335 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92b0e5d8_15a3_4889_a6f7_bb6ea79a2773.slice/crio-3a8e6bf8cd70c9968b54341b8ab3b78bf438eea85e3b0a386f3e5f952669a618 WatchSource:0}: Error finding container 3a8e6bf8cd70c9968b54341b8ab3b78bf438eea85e3b0a386f3e5f952669a618: Status 404 returned error can't find the container with id 3a8e6bf8cd70c9968b54341b8ab3b78bf438eea85e3b0a386f3e5f952669a618 Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.845295 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8skgc"] Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.846635 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.852649 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8skgc"] Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.936217 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jdcns"] Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.937660 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.942922 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jdcns"] Oct 04 04:01:37 crc kubenswrapper[4726]: I1004 04:01:37.967981 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sh9n\" (UniqueName: \"kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n\") pod \"nova-api-db-create-8skgc\" (UID: \"6ce587ac-9eb6-427d-b15c-399d40d440b9\") " pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.037519 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-x8ln8"] Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.038711 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.046987 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x8ln8"] Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.070135 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nr7c\" (UniqueName: \"kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c\") pod \"nova-cell0-db-create-jdcns\" (UID: \"a9830424-4983-4606-bf56-d6e0819d9eec\") " pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.070296 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sh9n\" (UniqueName: \"kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n\") pod \"nova-api-db-create-8skgc\" (UID: \"6ce587ac-9eb6-427d-b15c-399d40d440b9\") " pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.090543 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sh9n\" (UniqueName: \"kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n\") pod \"nova-api-db-create-8skgc\" (UID: \"6ce587ac-9eb6-427d-b15c-399d40d440b9\") " pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.171607 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nr7c\" (UniqueName: \"kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c\") pod \"nova-cell0-db-create-jdcns\" (UID: \"a9830424-4983-4606-bf56-d6e0819d9eec\") " pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.171793 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmhxb\" (UniqueName: \"kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb\") pod \"nova-cell1-db-create-x8ln8\" (UID: \"e70f6012-daec-42cb-b962-2864279cf632\") " pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.187626 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nr7c\" (UniqueName: \"kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c\") pod \"nova-cell0-db-create-jdcns\" (UID: \"a9830424-4983-4606-bf56-d6e0819d9eec\") " pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.198817 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.251352 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.273589 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmhxb\" (UniqueName: \"kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb\") pod \"nova-cell1-db-create-x8ln8\" (UID: \"e70f6012-daec-42cb-b962-2864279cf632\") " pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.290860 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmhxb\" (UniqueName: \"kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb\") pod \"nova-cell1-db-create-x8ln8\" (UID: \"e70f6012-daec-42cb-b962-2864279cf632\") " pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.352640 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.518213 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85cb2d95-8133-4c7f-8ee4-ca4683e39160" path="/var/lib/kubelet/pods/85cb2d95-8133-4c7f-8ee4-ca4683e39160/volumes" Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.690082 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8skgc"] Oct 04 04:01:38 crc kubenswrapper[4726]: W1004 04:01:38.691238 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ce587ac_9eb6_427d_b15c_399d40d440b9.slice/crio-afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548 WatchSource:0}: Error finding container afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548: Status 404 returned error can't find the container with id afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548 Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.740458 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerStarted","Data":"fa691922e0cf7669bdf62c6e4f6ea17a7b96a705e6e55345a28721d5a0002f76"} Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.740500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerStarted","Data":"3a8e6bf8cd70c9968b54341b8ab3b78bf438eea85e3b0a386f3e5f952669a618"} Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.742046 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8skgc" event={"ID":"6ce587ac-9eb6-427d-b15c-399d40d440b9","Type":"ContainerStarted","Data":"afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548"} Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.802351 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jdcns"] Oct 04 04:01:38 crc kubenswrapper[4726]: W1004 04:01:38.813541 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9830424_4983_4606_bf56_d6e0819d9eec.slice/crio-892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2 WatchSource:0}: Error finding container 892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2: Status 404 returned error can't find the container with id 892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2 Oct 04 04:01:38 crc kubenswrapper[4726]: I1004 04:01:38.924098 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-x8ln8"] Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.315601 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.315832 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.344910 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.373264 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.482699 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596368 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596466 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6dqb\" (UniqueName: \"kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596516 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596613 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596693 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596726 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.596803 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data\") pod \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\" (UID: \"5991ed37-a191-4fc6-b7cf-ff98a389b62b\") " Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.597822 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs" (OuterVolumeSpecName: "logs") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.602879 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb" (OuterVolumeSpecName: "kube-api-access-t6dqb") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "kube-api-access-t6dqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.604279 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.620484 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data" (OuterVolumeSpecName: "config-data") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.624730 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.634468 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts" (OuterVolumeSpecName: "scripts") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.650321 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "5991ed37-a191-4fc6-b7cf-ff98a389b62b" (UID: "5991ed37-a191-4fc6-b7cf-ff98a389b62b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.698950 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.698980 4726 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.698990 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.699000 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5991ed37-a191-4fc6-b7cf-ff98a389b62b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.699011 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6dqb\" (UniqueName: \"kubernetes.io/projected/5991ed37-a191-4fc6-b7cf-ff98a389b62b-kube-api-access-t6dqb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.699021 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5991ed37-a191-4fc6-b7cf-ff98a389b62b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.699029 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5991ed37-a191-4fc6-b7cf-ff98a389b62b-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.755681 4726 generic.go:334] "Generic (PLEG): container finished" podID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerID="3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6" exitCode=137 Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.755732 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerDied","Data":"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.755765 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b6cffb848-k9fxk" event={"ID":"5991ed37-a191-4fc6-b7cf-ff98a389b62b","Type":"ContainerDied","Data":"d32845fbbbd40ca9c1314f07f0f97f24428776b20c84b1f4666278a4c0beda45"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.755781 4726 scope.go:117] "RemoveContainer" containerID="267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.755873 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b6cffb848-k9fxk" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.762372 4726 generic.go:334] "Generic (PLEG): container finished" podID="e70f6012-daec-42cb-b962-2864279cf632" containerID="a2f202197b419f3f771ab7df93d402e98b1f68b10496add4d5d2e3d28bb8e634" exitCode=0 Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.762446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x8ln8" event={"ID":"e70f6012-daec-42cb-b962-2864279cf632","Type":"ContainerDied","Data":"a2f202197b419f3f771ab7df93d402e98b1f68b10496add4d5d2e3d28bb8e634"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.762473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x8ln8" event={"ID":"e70f6012-daec-42cb-b962-2864279cf632","Type":"ContainerStarted","Data":"1a3605bc3729c4d9caaf244afe8e59d6e38716858f48ecec93d8603f771e42ad"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.768838 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerStarted","Data":"91bfc80536631455a9a89ab4aace58ac01b06eabf34e89ade1389f70cf041511"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.770090 4726 generic.go:334] "Generic (PLEG): container finished" podID="6ce587ac-9eb6-427d-b15c-399d40d440b9" containerID="dec3a6a5787a134fadb9a0c8f2198464c5bd4617a6bf335d16b9dcb3c027af1b" exitCode=0 Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.770186 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8skgc" event={"ID":"6ce587ac-9eb6-427d-b15c-399d40d440b9","Type":"ContainerDied","Data":"dec3a6a5787a134fadb9a0c8f2198464c5bd4617a6bf335d16b9dcb3c027af1b"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.775570 4726 generic.go:334] "Generic (PLEG): container finished" podID="a9830424-4983-4606-bf56-d6e0819d9eec" containerID="2d49ea6ae564769a8f4f839243651683853779020afb3cb6050802db1fbd5c5c" exitCode=0 Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.776483 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jdcns" event={"ID":"a9830424-4983-4606-bf56-d6e0819d9eec","Type":"ContainerDied","Data":"2d49ea6ae564769a8f4f839243651683853779020afb3cb6050802db1fbd5c5c"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.776579 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jdcns" event={"ID":"a9830424-4983-4606-bf56-d6e0819d9eec","Type":"ContainerStarted","Data":"892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2"} Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.776609 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.776979 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.917799 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.925358 4726 scope.go:117] "RemoveContainer" containerID="3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.934590 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b6cffb848-k9fxk"] Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.959151 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.960097 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.962931 4726 scope.go:117] "RemoveContainer" containerID="267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4" Oct 04 04:01:39 crc kubenswrapper[4726]: E1004 04:01:39.963742 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4\": container with ID starting with 267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4 not found: ID does not exist" containerID="267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.963781 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4"} err="failed to get container status \"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4\": rpc error: code = NotFound desc = could not find container \"267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4\": container with ID starting with 267380ab1c2bc24b5841cf660530f96a281e3a1f5c266fdf722bbae6f6cc67b4 not found: ID does not exist" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.963803 4726 scope.go:117] "RemoveContainer" containerID="3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6" Oct 04 04:01:39 crc kubenswrapper[4726]: E1004 04:01:39.965058 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6\": container with ID starting with 3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6 not found: ID does not exist" containerID="3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.965096 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6"} err="failed to get container status \"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6\": rpc error: code = NotFound desc = could not find container \"3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6\": container with ID starting with 3f6f74979d02c02571034119cc92a22a245b8efb3a86f378e6e0eb0c9e1842e6 not found: ID does not exist" Oct 04 04:01:39 crc kubenswrapper[4726]: I1004 04:01:39.996013 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:01:40 crc kubenswrapper[4726]: I1004 04:01:40.028458 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:01:40 crc kubenswrapper[4726]: I1004 04:01:40.516908 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" path="/var/lib/kubelet/pods/5991ed37-a191-4fc6-b7cf-ff98a389b62b/volumes" Oct 04 04:01:40 crc kubenswrapper[4726]: I1004 04:01:40.785801 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerStarted","Data":"33a8c586fbf2a6a75b86d8ae7597290f41794040f86ae52c8cd5eb15c988a4e8"} Oct 04 04:01:40 crc kubenswrapper[4726]: I1004 04:01:40.788891 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:01:40 crc kubenswrapper[4726]: I1004 04:01:40.789065 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.189888 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.254685 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nr7c\" (UniqueName: \"kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c\") pod \"a9830424-4983-4606-bf56-d6e0819d9eec\" (UID: \"a9830424-4983-4606-bf56-d6e0819d9eec\") " Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.262747 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c" (OuterVolumeSpecName: "kube-api-access-6nr7c") pod "a9830424-4983-4606-bf56-d6e0819d9eec" (UID: "a9830424-4983-4606-bf56-d6e0819d9eec"). InnerVolumeSpecName "kube-api-access-6nr7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.332833 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.338221 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.355764 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmhxb\" (UniqueName: \"kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb\") pod \"e70f6012-daec-42cb-b962-2864279cf632\" (UID: \"e70f6012-daec-42cb-b962-2864279cf632\") " Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.355950 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sh9n\" (UniqueName: \"kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n\") pod \"6ce587ac-9eb6-427d-b15c-399d40d440b9\" (UID: \"6ce587ac-9eb6-427d-b15c-399d40d440b9\") " Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.356364 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nr7c\" (UniqueName: \"kubernetes.io/projected/a9830424-4983-4606-bf56-d6e0819d9eec-kube-api-access-6nr7c\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.359936 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n" (OuterVolumeSpecName: "kube-api-access-5sh9n") pod "6ce587ac-9eb6-427d-b15c-399d40d440b9" (UID: "6ce587ac-9eb6-427d-b15c-399d40d440b9"). InnerVolumeSpecName "kube-api-access-5sh9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.362304 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb" (OuterVolumeSpecName: "kube-api-access-nmhxb") pod "e70f6012-daec-42cb-b962-2864279cf632" (UID: "e70f6012-daec-42cb-b962-2864279cf632"). InnerVolumeSpecName "kube-api-access-nmhxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.458394 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmhxb\" (UniqueName: \"kubernetes.io/projected/e70f6012-daec-42cb-b962-2864279cf632-kube-api-access-nmhxb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.458806 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sh9n\" (UniqueName: \"kubernetes.io/projected/6ce587ac-9eb6-427d-b15c-399d40d440b9-kube-api-access-5sh9n\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.794987 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-x8ln8" event={"ID":"e70f6012-daec-42cb-b962-2864279cf632","Type":"ContainerDied","Data":"1a3605bc3729c4d9caaf244afe8e59d6e38716858f48ecec93d8603f771e42ad"} Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.795196 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a3605bc3729c4d9caaf244afe8e59d6e38716858f48ecec93d8603f771e42ad" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.795291 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-x8ln8" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.802264 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerStarted","Data":"68cfb1bc5b22fbc2d20e38be45f38b3ae558d7888d5a791ef9abb3508f06444f"} Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.802415 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.807778 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8skgc" event={"ID":"6ce587ac-9eb6-427d-b15c-399d40d440b9","Type":"ContainerDied","Data":"afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548"} Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.807826 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afcea48ba9ac936aa071a6d5be93aa6fb66dba62f6905344f5f3326f35077548" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.807793 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8skgc" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.809525 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jdcns" event={"ID":"a9830424-4983-4606-bf56-d6e0819d9eec","Type":"ContainerDied","Data":"892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2"} Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.809549 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.809566 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.809594 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jdcns" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.809554 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="892fb4836fbb7e48b0dc4dde2cc39fd9565581038c4eec27b187cd55825ad0b2" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.841723 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.843293 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:01:41 crc kubenswrapper[4726]: I1004 04:01:41.860817 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.160446814 podStartE2EDuration="5.860793543s" podCreationTimestamp="2025-10-04 04:01:36 +0000 UTC" firstStartedPulling="2025-10-04 04:01:37.712548829 +0000 UTC m=+1271.887172042" lastFinishedPulling="2025-10-04 04:01:41.412895568 +0000 UTC m=+1275.587518771" observedRunningTime="2025-10-04 04:01:41.828520075 +0000 UTC m=+1276.003143278" watchObservedRunningTime="2025-10-04 04:01:41.860793543 +0000 UTC m=+1276.035416756" Oct 04 04:01:42 crc kubenswrapper[4726]: I1004 04:01:42.816876 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:01:42 crc kubenswrapper[4726]: I1004 04:01:42.817124 4726 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:01:42 crc kubenswrapper[4726]: I1004 04:01:42.947205 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.402893 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.648989 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.827346 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-central-agent" containerID="cri-o://fa691922e0cf7669bdf62c6e4f6ea17a7b96a705e6e55345a28721d5a0002f76" gracePeriod=30 Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.827714 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="proxy-httpd" containerID="cri-o://68cfb1bc5b22fbc2d20e38be45f38b3ae558d7888d5a791ef9abb3508f06444f" gracePeriod=30 Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.827750 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="sg-core" containerID="cri-o://33a8c586fbf2a6a75b86d8ae7597290f41794040f86ae52c8cd5eb15c988a4e8" gracePeriod=30 Oct 04 04:01:43 crc kubenswrapper[4726]: I1004 04:01:43.827779 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-notification-agent" containerID="cri-o://91bfc80536631455a9a89ab4aace58ac01b06eabf34e89ade1389f70cf041511" gracePeriod=30 Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.837921 4726 generic.go:334] "Generic (PLEG): container finished" podID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerID="68cfb1bc5b22fbc2d20e38be45f38b3ae558d7888d5a791ef9abb3508f06444f" exitCode=0 Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.837990 4726 generic.go:334] "Generic (PLEG): container finished" podID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerID="33a8c586fbf2a6a75b86d8ae7597290f41794040f86ae52c8cd5eb15c988a4e8" exitCode=2 Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.838011 4726 generic.go:334] "Generic (PLEG): container finished" podID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerID="91bfc80536631455a9a89ab4aace58ac01b06eabf34e89ade1389f70cf041511" exitCode=0 Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.838021 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerDied","Data":"68cfb1bc5b22fbc2d20e38be45f38b3ae558d7888d5a791ef9abb3508f06444f"} Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.838085 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerDied","Data":"33a8c586fbf2a6a75b86d8ae7597290f41794040f86ae52c8cd5eb15c988a4e8"} Oct 04 04:01:44 crc kubenswrapper[4726]: I1004 04:01:44.838101 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerDied","Data":"91bfc80536631455a9a89ab4aace58ac01b06eabf34e89ade1389f70cf041511"} Oct 04 04:01:46 crc kubenswrapper[4726]: I1004 04:01:46.887518 4726 generic.go:334] "Generic (PLEG): container finished" podID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerID="fa691922e0cf7669bdf62c6e4f6ea17a7b96a705e6e55345a28721d5a0002f76" exitCode=0 Oct 04 04:01:46 crc kubenswrapper[4726]: I1004 04:01:46.887589 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerDied","Data":"fa691922e0cf7669bdf62c6e4f6ea17a7b96a705e6e55345a28721d5a0002f76"} Oct 04 04:01:46 crc kubenswrapper[4726]: I1004 04:01:46.947524 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078489 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078609 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078652 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078733 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078750 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078773 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxqhb\" (UniqueName: \"kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.078824 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle\") pod \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\" (UID: \"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773\") " Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.081410 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.081665 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.102341 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts" (OuterVolumeSpecName: "scripts") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.102444 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb" (OuterVolumeSpecName: "kube-api-access-bxqhb") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "kube-api-access-bxqhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.128276 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.175379 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181221 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181268 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181279 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181290 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181300 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.181309 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxqhb\" (UniqueName: \"kubernetes.io/projected/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-kube-api-access-bxqhb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.214238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data" (OuterVolumeSpecName: "config-data") pod "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" (UID: "92b0e5d8-15a3-4889-a6f7-bb6ea79a2773"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.283558 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.900398 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92b0e5d8-15a3-4889-a6f7-bb6ea79a2773","Type":"ContainerDied","Data":"3a8e6bf8cd70c9968b54341b8ab3b78bf438eea85e3b0a386f3e5f952669a618"} Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.900516 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.901603 4726 scope.go:117] "RemoveContainer" containerID="68cfb1bc5b22fbc2d20e38be45f38b3ae558d7888d5a791ef9abb3508f06444f" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.938358 4726 scope.go:117] "RemoveContainer" containerID="33a8c586fbf2a6a75b86d8ae7597290f41794040f86ae52c8cd5eb15c988a4e8" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.950619 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.960146 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.969219 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3795-account-create-b95q8"] Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.975657 4726 scope.go:117] "RemoveContainer" containerID="91bfc80536631455a9a89ab4aace58ac01b06eabf34e89ade1389f70cf041511" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976427 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="sg-core" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976461 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="sg-core" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976481 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="proxy-httpd" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976488 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="proxy-httpd" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976503 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ce587ac-9eb6-427d-b15c-399d40d440b9" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976509 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ce587ac-9eb6-427d-b15c-399d40d440b9" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976520 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976526 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976539 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70f6012-daec-42cb-b962-2864279cf632" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976545 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70f6012-daec-42cb-b962-2864279cf632" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976554 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-notification-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976559 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-notification-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976578 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9830424-4983-4606-bf56-d6e0819d9eec" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976584 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9830424-4983-4606-bf56-d6e0819d9eec" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976595 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-central-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976601 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-central-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: E1004 04:01:47.976611 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon-log" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976617 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon-log" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976792 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9830424-4983-4606-bf56-d6e0819d9eec" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976800 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon-log" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976814 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-central-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976826 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ce587ac-9eb6-427d-b15c-399d40d440b9" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976835 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="ceilometer-notification-agent" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976847 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="proxy-httpd" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976859 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70f6012-daec-42cb-b962-2864279cf632" containerName="mariadb-database-create" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976870 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" containerName="sg-core" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.976877 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5991ed37-a191-4fc6-b7cf-ff98a389b62b" containerName="horizon" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.977500 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.977966 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3795-account-create-b95q8"] Oct 04 04:01:47 crc kubenswrapper[4726]: I1004 04:01:47.980301 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.009202 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.010847 4726 scope.go:117] "RemoveContainer" containerID="fa691922e0cf7669bdf62c6e4f6ea17a7b96a705e6e55345a28721d5a0002f76" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.011990 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.015500 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.020014 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.035660 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102167 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102236 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102277 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vgnn\" (UniqueName: \"kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102330 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102357 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102445 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102481 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xsl\" (UniqueName: \"kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl\") pod \"nova-api-3795-account-create-b95q8\" (UID: \"200b106d-3b99-40c1-9368-32370f56d01a\") " pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.102512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.119974 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-d3fb-account-create-7v5gc"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.121246 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.122872 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.130805 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d3fb-account-create-7v5gc"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.206733 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.206827 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.206875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w8dl\" (UniqueName: \"kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl\") pod \"nova-cell0-d3fb-account-create-7v5gc\" (UID: \"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69\") " pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.206919 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xsl\" (UniqueName: \"kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl\") pod \"nova-api-3795-account-create-b95q8\" (UID: \"200b106d-3b99-40c1-9368-32370f56d01a\") " pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.206957 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207018 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207062 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207097 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vgnn\" (UniqueName: \"kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207174 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.207729 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.210969 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.211867 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.212068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.214089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.229022 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vgnn\" (UniqueName: \"kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn\") pod \"ceilometer-0\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.229142 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xsl\" (UniqueName: \"kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl\") pod \"nova-api-3795-account-create-b95q8\" (UID: \"200b106d-3b99-40c1-9368-32370f56d01a\") " pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.300671 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.308946 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w8dl\" (UniqueName: \"kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl\") pod \"nova-cell0-d3fb-account-create-7v5gc\" (UID: \"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69\") " pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.331733 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w8dl\" (UniqueName: \"kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl\") pod \"nova-cell0-d3fb-account-create-7v5gc\" (UID: \"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69\") " pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.332017 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.336151 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-54bd-account-create-698qk"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.337325 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.339432 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.346169 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54bd-account-create-698qk"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.439607 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.512577 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2ltb\" (UniqueName: \"kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb\") pod \"nova-cell1-54bd-account-create-698qk\" (UID: \"b65ae8b8-8121-45de-9647-f90d71a923c9\") " pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.515194 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92b0e5d8-15a3-4889-a6f7-bb6ea79a2773" path="/var/lib/kubelet/pods/92b0e5d8-15a3-4889-a6f7-bb6ea79a2773/volumes" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.614581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2ltb\" (UniqueName: \"kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb\") pod \"nova-cell1-54bd-account-create-698qk\" (UID: \"b65ae8b8-8121-45de-9647-f90d71a923c9\") " pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.647893 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2ltb\" (UniqueName: \"kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb\") pod \"nova-cell1-54bd-account-create-698qk\" (UID: \"b65ae8b8-8121-45de-9647-f90d71a923c9\") " pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.740758 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.848003 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3795-account-create-b95q8"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.863280 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.912259 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerStarted","Data":"6c16c9a07f4e58815d498254a5d3d7194d2faa4d6c1ba027b3a5703ac6ee6044"} Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.917152 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3795-account-create-b95q8" event={"ID":"200b106d-3b99-40c1-9368-32370f56d01a","Type":"ContainerStarted","Data":"ad2de024249352ab0f101884440d0eb1ed9e47f79a7cfd65597560ba6e0dae73"} Oct 04 04:01:48 crc kubenswrapper[4726]: I1004 04:01:48.991347 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d3fb-account-create-7v5gc"] Oct 04 04:01:48 crc kubenswrapper[4726]: W1004 04:01:48.993373 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e2a19f3_d51c_4ea0_a16c_3b0bad069a69.slice/crio-bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9 WatchSource:0}: Error finding container bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9: Status 404 returned error can't find the container with id bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9 Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.210742 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-54bd-account-create-698qk"] Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.935705 4726 generic.go:334] "Generic (PLEG): container finished" podID="7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" containerID="c71c59d33610947f261bf22c3e0bd3659db271acc042bdef71c860d35774e971" exitCode=0 Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.935755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" event={"ID":"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69","Type":"ContainerDied","Data":"c71c59d33610947f261bf22c3e0bd3659db271acc042bdef71c860d35774e971"} Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.935797 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" event={"ID":"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69","Type":"ContainerStarted","Data":"bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9"} Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.938562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54bd-account-create-698qk" event={"ID":"b65ae8b8-8121-45de-9647-f90d71a923c9","Type":"ContainerStarted","Data":"4267fbb25d94da276ddf0ffce8beaae1b0e747cae4297258c39dfba7eb471dbd"} Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.946390 4726 generic.go:334] "Generic (PLEG): container finished" podID="200b106d-3b99-40c1-9368-32370f56d01a" containerID="8052916d2328c64bfe226e9102cc5b3b18c541bec7577f8fbe032b05cbc26c31" exitCode=0 Oct 04 04:01:49 crc kubenswrapper[4726]: I1004 04:01:49.946441 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3795-account-create-b95q8" event={"ID":"200b106d-3b99-40c1-9368-32370f56d01a","Type":"ContainerDied","Data":"8052916d2328c64bfe226e9102cc5b3b18c541bec7577f8fbe032b05cbc26c31"} Oct 04 04:01:50 crc kubenswrapper[4726]: I1004 04:01:50.958153 4726 generic.go:334] "Generic (PLEG): container finished" podID="b65ae8b8-8121-45de-9647-f90d71a923c9" containerID="f1b3c30f91cca95b2b19725869571981aecb3fb57d2281b1f9d07e7414ea2ed7" exitCode=0 Oct 04 04:01:50 crc kubenswrapper[4726]: I1004 04:01:50.958239 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54bd-account-create-698qk" event={"ID":"b65ae8b8-8121-45de-9647-f90d71a923c9","Type":"ContainerDied","Data":"f1b3c30f91cca95b2b19725869571981aecb3fb57d2281b1f9d07e7414ea2ed7"} Oct 04 04:01:50 crc kubenswrapper[4726]: I1004 04:01:50.962161 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerStarted","Data":"75b3de1777901cc8b11d95c0fd1a50b888ab35b42f683591354e52c0c10b5a79"} Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.418064 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.421713 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.576867 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2xsl\" (UniqueName: \"kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl\") pod \"200b106d-3b99-40c1-9368-32370f56d01a\" (UID: \"200b106d-3b99-40c1-9368-32370f56d01a\") " Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.576915 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w8dl\" (UniqueName: \"kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl\") pod \"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69\" (UID: \"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69\") " Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.582847 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl" (OuterVolumeSpecName: "kube-api-access-5w8dl") pod "7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" (UID: "7e2a19f3-d51c-4ea0-a16c-3b0bad069a69"). InnerVolumeSpecName "kube-api-access-5w8dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.584317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl" (OuterVolumeSpecName: "kube-api-access-s2xsl") pod "200b106d-3b99-40c1-9368-32370f56d01a" (UID: "200b106d-3b99-40c1-9368-32370f56d01a"). InnerVolumeSpecName "kube-api-access-s2xsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.679938 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2xsl\" (UniqueName: \"kubernetes.io/projected/200b106d-3b99-40c1-9368-32370f56d01a-kube-api-access-s2xsl\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.680179 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w8dl\" (UniqueName: \"kubernetes.io/projected/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69-kube-api-access-5w8dl\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.973064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" event={"ID":"7e2a19f3-d51c-4ea0-a16c-3b0bad069a69","Type":"ContainerDied","Data":"bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9"} Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.974245 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc715fdc4bff236c6fa1dca97bfe31ae28459d8f7120c00d9e4f55939e2696d9" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.973254 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d3fb-account-create-7v5gc" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.975399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerStarted","Data":"8b5ca6873b8feac320263ba565ea94ee21d2727b63063b94aa4d6336ddcbe96c"} Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.975440 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerStarted","Data":"199b223350a752fb1aa7c9d8aeb77bb13c60ea1e2099c61de86e758795d3cffa"} Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.977447 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3795-account-create-b95q8" event={"ID":"200b106d-3b99-40c1-9368-32370f56d01a","Type":"ContainerDied","Data":"ad2de024249352ab0f101884440d0eb1ed9e47f79a7cfd65597560ba6e0dae73"} Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.977506 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad2de024249352ab0f101884440d0eb1ed9e47f79a7cfd65597560ba6e0dae73" Oct 04 04:01:51 crc kubenswrapper[4726]: I1004 04:01:51.977549 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3795-account-create-b95q8" Oct 04 04:01:52 crc kubenswrapper[4726]: I1004 04:01:52.326804 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:52 crc kubenswrapper[4726]: I1004 04:01:52.495573 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2ltb\" (UniqueName: \"kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb\") pod \"b65ae8b8-8121-45de-9647-f90d71a923c9\" (UID: \"b65ae8b8-8121-45de-9647-f90d71a923c9\") " Oct 04 04:01:52 crc kubenswrapper[4726]: I1004 04:01:52.508365 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb" (OuterVolumeSpecName: "kube-api-access-b2ltb") pod "b65ae8b8-8121-45de-9647-f90d71a923c9" (UID: "b65ae8b8-8121-45de-9647-f90d71a923c9"). InnerVolumeSpecName "kube-api-access-b2ltb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:01:52 crc kubenswrapper[4726]: I1004 04:01:52.600850 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2ltb\" (UniqueName: \"kubernetes.io/projected/b65ae8b8-8121-45de-9647-f90d71a923c9-kube-api-access-b2ltb\") on node \"crc\" DevicePath \"\"" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.000204 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-54bd-account-create-698qk" event={"ID":"b65ae8b8-8121-45de-9647-f90d71a923c9","Type":"ContainerDied","Data":"4267fbb25d94da276ddf0ffce8beaae1b0e747cae4297258c39dfba7eb471dbd"} Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.000249 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4267fbb25d94da276ddf0ffce8beaae1b0e747cae4297258c39dfba7eb471dbd" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.000339 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-54bd-account-create-698qk" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.493481 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7t6dt"] Oct 04 04:01:53 crc kubenswrapper[4726]: E1004 04:01:53.494336 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494363 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: E1004 04:01:53.494392 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b65ae8b8-8121-45de-9647-f90d71a923c9" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494399 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b65ae8b8-8121-45de-9647-f90d71a923c9" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: E1004 04:01:53.494413 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200b106d-3b99-40c1-9368-32370f56d01a" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494419 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="200b106d-3b99-40c1-9368-32370f56d01a" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494582 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="200b106d-3b99-40c1-9368-32370f56d01a" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494597 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b65ae8b8-8121-45de-9647-f90d71a923c9" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.494610 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" containerName="mariadb-account-create" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.495200 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.502694 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.502746 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.502815 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fxc5g" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.508366 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7t6dt"] Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.618033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.618062 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.618091 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbnd9\" (UniqueName: \"kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.618206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.719714 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.719823 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.719843 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.719868 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbnd9\" (UniqueName: \"kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.724656 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.737644 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.749598 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbnd9\" (UniqueName: \"kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.749713 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7t6dt\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:53 crc kubenswrapper[4726]: I1004 04:01:53.857867 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:01:54 crc kubenswrapper[4726]: I1004 04:01:54.018757 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerStarted","Data":"607a51c47b05ca9bd363accffee03bceebcfb02d46cdbb9580907d06ce4c17ec"} Oct 04 04:01:54 crc kubenswrapper[4726]: I1004 04:01:54.018973 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:01:54 crc kubenswrapper[4726]: I1004 04:01:54.047947 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.620745514 podStartE2EDuration="7.047925526s" podCreationTimestamp="2025-10-04 04:01:47 +0000 UTC" firstStartedPulling="2025-10-04 04:01:48.891961708 +0000 UTC m=+1283.066584911" lastFinishedPulling="2025-10-04 04:01:53.31914171 +0000 UTC m=+1287.493764923" observedRunningTime="2025-10-04 04:01:54.037850216 +0000 UTC m=+1288.212473439" watchObservedRunningTime="2025-10-04 04:01:54.047925526 +0000 UTC m=+1288.222548739" Oct 04 04:01:54 crc kubenswrapper[4726]: I1004 04:01:54.305518 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7t6dt"] Oct 04 04:01:54 crc kubenswrapper[4726]: W1004 04:01:54.307078 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88100d2e_2e91_49d5_8f3f_f51864942282.slice/crio-812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3 WatchSource:0}: Error finding container 812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3: Status 404 returned error can't find the container with id 812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3 Oct 04 04:01:55 crc kubenswrapper[4726]: I1004 04:01:55.036248 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" event={"ID":"88100d2e-2e91-49d5-8f3f-f51864942282","Type":"ContainerStarted","Data":"812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3"} Oct 04 04:01:59 crc kubenswrapper[4726]: I1004 04:01:59.916315 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:01:59 crc kubenswrapper[4726]: I1004 04:01:59.917096 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-central-agent" containerID="cri-o://75b3de1777901cc8b11d95c0fd1a50b888ab35b42f683591354e52c0c10b5a79" gracePeriod=30 Oct 04 04:01:59 crc kubenswrapper[4726]: I1004 04:01:59.917510 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="proxy-httpd" containerID="cri-o://607a51c47b05ca9bd363accffee03bceebcfb02d46cdbb9580907d06ce4c17ec" gracePeriod=30 Oct 04 04:01:59 crc kubenswrapper[4726]: I1004 04:01:59.917560 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="sg-core" containerID="cri-o://8b5ca6873b8feac320263ba565ea94ee21d2727b63063b94aa4d6336ddcbe96c" gracePeriod=30 Oct 04 04:01:59 crc kubenswrapper[4726]: I1004 04:01:59.917593 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-notification-agent" containerID="cri-o://199b223350a752fb1aa7c9d8aeb77bb13c60ea1e2099c61de86e758795d3cffa" gracePeriod=30 Oct 04 04:02:00 crc kubenswrapper[4726]: I1004 04:02:00.087660 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerID="607a51c47b05ca9bd363accffee03bceebcfb02d46cdbb9580907d06ce4c17ec" exitCode=0 Oct 04 04:02:00 crc kubenswrapper[4726]: I1004 04:02:00.087930 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerID="8b5ca6873b8feac320263ba565ea94ee21d2727b63063b94aa4d6336ddcbe96c" exitCode=2 Oct 04 04:02:00 crc kubenswrapper[4726]: I1004 04:02:00.087730 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerDied","Data":"607a51c47b05ca9bd363accffee03bceebcfb02d46cdbb9580907d06ce4c17ec"} Oct 04 04:02:00 crc kubenswrapper[4726]: I1004 04:02:00.087966 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerDied","Data":"8b5ca6873b8feac320263ba565ea94ee21d2727b63063b94aa4d6336ddcbe96c"} Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.101701 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerID="199b223350a752fb1aa7c9d8aeb77bb13c60ea1e2099c61de86e758795d3cffa" exitCode=0 Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.101736 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerID="75b3de1777901cc8b11d95c0fd1a50b888ab35b42f683591354e52c0c10b5a79" exitCode=0 Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.101752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerDied","Data":"199b223350a752fb1aa7c9d8aeb77bb13c60ea1e2099c61de86e758795d3cffa"} Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.101775 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerDied","Data":"75b3de1777901cc8b11d95c0fd1a50b888ab35b42f683591354e52c0c10b5a79"} Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.515239 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.665905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666016 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666042 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666143 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666206 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666225 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vgnn\" (UniqueName: \"kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666264 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml\") pod \"ac326902-f4df-4f97-bd4a-8424f01dbced\" (UID: \"ac326902-f4df-4f97-bd4a-8424f01dbced\") " Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.666873 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.667036 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.672917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn" (OuterVolumeSpecName: "kube-api-access-9vgnn") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "kube-api-access-9vgnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.674073 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts" (OuterVolumeSpecName: "scripts") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.692239 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.772898 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.772933 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.772946 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac326902-f4df-4f97-bd4a-8424f01dbced-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.772957 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vgnn\" (UniqueName: \"kubernetes.io/projected/ac326902-f4df-4f97-bd4a-8424f01dbced-kube-api-access-9vgnn\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.772972 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.775249 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.795268 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data" (OuterVolumeSpecName: "config-data") pod "ac326902-f4df-4f97-bd4a-8424f01dbced" (UID: "ac326902-f4df-4f97-bd4a-8424f01dbced"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.875230 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:01 crc kubenswrapper[4726]: I1004 04:02:01.875453 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac326902-f4df-4f97-bd4a-8424f01dbced-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.119277 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac326902-f4df-4f97-bd4a-8424f01dbced","Type":"ContainerDied","Data":"6c16c9a07f4e58815d498254a5d3d7194d2faa4d6c1ba027b3a5703ac6ee6044"} Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.120836 4726 scope.go:117] "RemoveContainer" containerID="607a51c47b05ca9bd363accffee03bceebcfb02d46cdbb9580907d06ce4c17ec" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.121503 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.130076 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" event={"ID":"88100d2e-2e91-49d5-8f3f-f51864942282","Type":"ContainerStarted","Data":"8acbc066cb63c6b186f8b40efaade29ed567885912707f43633abe27f145e77d"} Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.167393 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" podStartSLOduration=2.221026292 podStartE2EDuration="9.16735671s" podCreationTimestamp="2025-10-04 04:01:53 +0000 UTC" firstStartedPulling="2025-10-04 04:01:54.309248673 +0000 UTC m=+1288.483871886" lastFinishedPulling="2025-10-04 04:02:01.255579071 +0000 UTC m=+1295.430202304" observedRunningTime="2025-10-04 04:02:02.150798194 +0000 UTC m=+1296.325421417" watchObservedRunningTime="2025-10-04 04:02:02.16735671 +0000 UTC m=+1296.341979993" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.167737 4726 scope.go:117] "RemoveContainer" containerID="8b5ca6873b8feac320263ba565ea94ee21d2727b63063b94aa4d6336ddcbe96c" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.193570 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.200902 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.214446 4726 scope.go:117] "RemoveContainer" containerID="199b223350a752fb1aa7c9d8aeb77bb13c60ea1e2099c61de86e758795d3cffa" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221146 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:02 crc kubenswrapper[4726]: E1004 04:02:02.221557 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-central-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221582 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-central-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: E1004 04:02:02.221613 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="proxy-httpd" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221621 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="proxy-httpd" Oct 04 04:02:02 crc kubenswrapper[4726]: E1004 04:02:02.221642 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="sg-core" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221649 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="sg-core" Oct 04 04:02:02 crc kubenswrapper[4726]: E1004 04:02:02.221662 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-notification-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221692 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-notification-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221900 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="sg-core" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221914 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-notification-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221926 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="proxy-httpd" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.221945 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" containerName="ceilometer-central-agent" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.225263 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.227060 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.227263 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.240032 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.268515 4726 scope.go:117] "RemoveContainer" containerID="75b3de1777901cc8b11d95c0fd1a50b888ab35b42f683591354e52c0c10b5a79" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.282941 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.282982 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.283001 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.283021 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.283044 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zns99\" (UniqueName: \"kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.283084 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.283121 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384508 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384557 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384581 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zns99\" (UniqueName: \"kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384693 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.384719 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.385571 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.385663 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.389461 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.389854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.394346 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.402432 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.415590 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zns99\" (UniqueName: \"kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99\") pod \"ceilometer-0\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " pod="openstack/ceilometer-0" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.517485 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac326902-f4df-4f97-bd4a-8424f01dbced" path="/var/lib/kubelet/pods/ac326902-f4df-4f97-bd4a-8424f01dbced/volumes" Oct 04 04:02:02 crc kubenswrapper[4726]: I1004 04:02:02.560673 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:03 crc kubenswrapper[4726]: I1004 04:02:03.012405 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:03 crc kubenswrapper[4726]: W1004 04:02:03.016395 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb2708ed_c9c1_4d6d_83ab_d48c621ed688.slice/crio-563467c6512879817f649766b69c8ce7b5bd4e8f304b05cdd1c1c26252647877 WatchSource:0}: Error finding container 563467c6512879817f649766b69c8ce7b5bd4e8f304b05cdd1c1c26252647877: Status 404 returned error can't find the container with id 563467c6512879817f649766b69c8ce7b5bd4e8f304b05cdd1c1c26252647877 Oct 04 04:02:03 crc kubenswrapper[4726]: I1004 04:02:03.161068 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerStarted","Data":"563467c6512879817f649766b69c8ce7b5bd4e8f304b05cdd1c1c26252647877"} Oct 04 04:02:04 crc kubenswrapper[4726]: I1004 04:02:04.176314 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerStarted","Data":"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914"} Oct 04 04:02:05 crc kubenswrapper[4726]: I1004 04:02:05.194742 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerStarted","Data":"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e"} Oct 04 04:02:06 crc kubenswrapper[4726]: I1004 04:02:06.205686 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerStarted","Data":"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d"} Oct 04 04:02:08 crc kubenswrapper[4726]: I1004 04:02:08.237881 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerStarted","Data":"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63"} Oct 04 04:02:08 crc kubenswrapper[4726]: I1004 04:02:08.239233 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:02:08 crc kubenswrapper[4726]: I1004 04:02:08.282188 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.2518147969999998 podStartE2EDuration="6.282073831s" podCreationTimestamp="2025-10-04 04:02:02 +0000 UTC" firstStartedPulling="2025-10-04 04:02:03.019726501 +0000 UTC m=+1297.194349714" lastFinishedPulling="2025-10-04 04:02:07.049985495 +0000 UTC m=+1301.224608748" observedRunningTime="2025-10-04 04:02:08.266521223 +0000 UTC m=+1302.441144476" watchObservedRunningTime="2025-10-04 04:02:08.282073831 +0000 UTC m=+1302.456697064" Oct 04 04:02:13 crc kubenswrapper[4726]: I1004 04:02:13.306444 4726 generic.go:334] "Generic (PLEG): container finished" podID="88100d2e-2e91-49d5-8f3f-f51864942282" containerID="8acbc066cb63c6b186f8b40efaade29ed567885912707f43633abe27f145e77d" exitCode=0 Oct 04 04:02:13 crc kubenswrapper[4726]: I1004 04:02:13.306587 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" event={"ID":"88100d2e-2e91-49d5-8f3f-f51864942282","Type":"ContainerDied","Data":"8acbc066cb63c6b186f8b40efaade29ed567885912707f43633abe27f145e77d"} Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.811377 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.836279 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbnd9\" (UniqueName: \"kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9\") pod \"88100d2e-2e91-49d5-8f3f-f51864942282\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.836321 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle\") pod \"88100d2e-2e91-49d5-8f3f-f51864942282\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.836428 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts\") pod \"88100d2e-2e91-49d5-8f3f-f51864942282\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.836600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data\") pod \"88100d2e-2e91-49d5-8f3f-f51864942282\" (UID: \"88100d2e-2e91-49d5-8f3f-f51864942282\") " Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.844079 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9" (OuterVolumeSpecName: "kube-api-access-xbnd9") pod "88100d2e-2e91-49d5-8f3f-f51864942282" (UID: "88100d2e-2e91-49d5-8f3f-f51864942282"). InnerVolumeSpecName "kube-api-access-xbnd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.845553 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts" (OuterVolumeSpecName: "scripts") pod "88100d2e-2e91-49d5-8f3f-f51864942282" (UID: "88100d2e-2e91-49d5-8f3f-f51864942282"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.871388 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data" (OuterVolumeSpecName: "config-data") pod "88100d2e-2e91-49d5-8f3f-f51864942282" (UID: "88100d2e-2e91-49d5-8f3f-f51864942282"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.890642 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88100d2e-2e91-49d5-8f3f-f51864942282" (UID: "88100d2e-2e91-49d5-8f3f-f51864942282"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.938215 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.938256 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbnd9\" (UniqueName: \"kubernetes.io/projected/88100d2e-2e91-49d5-8f3f-f51864942282-kube-api-access-xbnd9\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.938275 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:14 crc kubenswrapper[4726]: I1004 04:02:14.938292 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88100d2e-2e91-49d5-8f3f-f51864942282-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.328813 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" event={"ID":"88100d2e-2e91-49d5-8f3f-f51864942282","Type":"ContainerDied","Data":"812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3"} Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.329351 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="812edabff0cdeb5694d1728071b68a8a07c4d6887b6fa15394baf832e0fe27e3" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.328996 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7t6dt" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.449494 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:02:15 crc kubenswrapper[4726]: E1004 04:02:15.449892 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88100d2e-2e91-49d5-8f3f-f51864942282" containerName="nova-cell0-conductor-db-sync" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.449914 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="88100d2e-2e91-49d5-8f3f-f51864942282" containerName="nova-cell0-conductor-db-sync" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.451231 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="88100d2e-2e91-49d5-8f3f-f51864942282" containerName="nova-cell0-conductor-db-sync" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.452319 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.456796 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fxc5g" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.458052 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.485922 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.553880 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x57wt\" (UniqueName: \"kubernetes.io/projected/76642298-36b4-4bde-ba49-3b0a2e56494d-kube-api-access-x57wt\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.553937 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.554039 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.655158 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x57wt\" (UniqueName: \"kubernetes.io/projected/76642298-36b4-4bde-ba49-3b0a2e56494d-kube-api-access-x57wt\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.655436 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.655587 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.661423 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.662599 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76642298-36b4-4bde-ba49-3b0a2e56494d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.673248 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x57wt\" (UniqueName: \"kubernetes.io/projected/76642298-36b4-4bde-ba49-3b0a2e56494d-kube-api-access-x57wt\") pod \"nova-cell0-conductor-0\" (UID: \"76642298-36b4-4bde-ba49-3b0a2e56494d\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:15 crc kubenswrapper[4726]: I1004 04:02:15.799624 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:16 crc kubenswrapper[4726]: I1004 04:02:16.328783 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:02:17 crc kubenswrapper[4726]: I1004 04:02:17.355677 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76642298-36b4-4bde-ba49-3b0a2e56494d","Type":"ContainerStarted","Data":"9611230dd3e91de133865f0d8b64657c09bd71191caaf021c3a60c845f14d364"} Oct 04 04:02:17 crc kubenswrapper[4726]: I1004 04:02:17.356033 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76642298-36b4-4bde-ba49-3b0a2e56494d","Type":"ContainerStarted","Data":"6880589adfa80c10a857bf5e62b76b605ded71f7fd6647cc61628ef6bac9678f"} Oct 04 04:02:17 crc kubenswrapper[4726]: I1004 04:02:17.356433 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:17 crc kubenswrapper[4726]: I1004 04:02:17.385346 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.385317596 podStartE2EDuration="2.385317596s" podCreationTimestamp="2025-10-04 04:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:17.376835782 +0000 UTC m=+1311.551458995" watchObservedRunningTime="2025-10-04 04:02:17.385317596 +0000 UTC m=+1311.559940849" Oct 04 04:02:25 crc kubenswrapper[4726]: I1004 04:02:25.844838 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.353840 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-m8rwl"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.355772 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.357909 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.358219 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.365544 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m8rwl"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.454829 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt6m4\" (UniqueName: \"kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.454903 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.455000 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.455194 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.496741 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.499883 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.521387 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.537198 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564283 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564336 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564403 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564484 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt6m4\" (UniqueName: \"kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564510 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564562 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.564606 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcs2f\" (UniqueName: \"kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.571918 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.572163 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.589965 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.589988 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.601083 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.613954 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt6m4\" (UniqueName: \"kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4\") pod \"nova-cell0-cell-mapping-m8rwl\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.614028 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.615460 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.625895 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.633395 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.634671 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.635277 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.644691 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.648162 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666142 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvzcg\" (UniqueName: \"kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666215 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666277 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666319 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krp7t\" (UniqueName: \"kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666357 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcs2f\" (UniqueName: \"kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666405 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666435 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666472 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666503 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.666545 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.667000 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.674374 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.687874 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.701073 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcs2f\" (UniqueName: \"kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f\") pod \"nova-api-0\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.701476 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.764349 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.768048 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.769195 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvzcg\" (UniqueName: \"kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.769241 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.769290 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.770319 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krp7t\" (UniqueName: \"kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.770347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.770423 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.770462 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.771543 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.776452 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.780835 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.780900 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.781743 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.789010 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krp7t\" (UniqueName: \"kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.792068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvzcg\" (UniqueName: \"kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg\") pod \"nova-metadata-0\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.792727 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.809583 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.815941 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.817487 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.828404 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.848870 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.858937 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.860642 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.885689 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.885730 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z674z\" (UniqueName: \"kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.888519 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.888588 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.888618 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.888633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4gsd\" (UniqueName: \"kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991822 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991860 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991880 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z674z\" (UniqueName: \"kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991924 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991956 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.991981 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.992001 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.992072 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.993021 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.993230 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.993232 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.993325 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:26 crc kubenswrapper[4726]: I1004 04:02:26.993535 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.022879 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z674z\" (UniqueName: \"kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z\") pod \"dnsmasq-dns-865f5d856f-9nfgv\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.094504 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4gsd\" (UniqueName: \"kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.094595 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.094705 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.098272 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.101301 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.115810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4gsd\" (UniqueName: \"kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd\") pod \"nova-scheduler-0\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.169260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.172066 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.370546 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m8rwl"] Oct 04 04:02:27 crc kubenswrapper[4726]: W1004 04:02:27.483618 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55722bd_a894_4195_b302_1a56c45b8579.slice/crio-f537c14de8809f86bd412ddb10d24b00787cb51f9e801d4ead99fc31dd9c2e27 WatchSource:0}: Error finding container f537c14de8809f86bd412ddb10d24b00787cb51f9e801d4ead99fc31dd9c2e27: Status 404 returned error can't find the container with id f537c14de8809f86bd412ddb10d24b00787cb51f9e801d4ead99fc31dd9c2e27 Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.484349 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.503916 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x69lg"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.505745 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.513689 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.513891 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.520492 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x69lg"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.528445 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m8rwl" event={"ID":"b3dba01d-46b8-407b-9622-5fd1534b9edc","Type":"ContainerStarted","Data":"31637f9dd1d67eaf73e05aa3ece1f84f39e8adc4904f1418b812ea0e9c05fd95"} Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.530183 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerStarted","Data":"f537c14de8809f86bd412ddb10d24b00787cb51f9e801d4ead99fc31dd9c2e27"} Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.582021 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:02:27 crc kubenswrapper[4726]: W1004 04:02:27.588564 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded51a4d9_f248_470f_a887_1ae137e39a66.slice/crio-057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a WatchSource:0}: Error finding container 057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a: Status 404 returned error can't find the container with id 057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.590080 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.605413 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhmv\" (UniqueName: \"kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.605457 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.605536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.605658 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.707787 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhmv\" (UniqueName: \"kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.707829 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.707903 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.707951 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.713283 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.715691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.715714 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.727373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhmv\" (UniqueName: \"kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv\") pod \"nova-cell1-conductor-db-sync-x69lg\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.800993 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.813711 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:27 crc kubenswrapper[4726]: I1004 04:02:27.859633 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.464098 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x69lg"] Oct 04 04:02:28 crc kubenswrapper[4726]: W1004 04:02:28.473316 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e8ff110_c94f_4dc0_8420_e00cff505a20.slice/crio-ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55 WatchSource:0}: Error finding container ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55: Status 404 returned error can't find the container with id ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55 Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.541149 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed51a4d9-f248-470f-a887-1ae137e39a66","Type":"ContainerStarted","Data":"057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.543648 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerStarted","Data":"ea0d45a51987715d2693de0f361553d3e99255e63a720fb97e309c375cf1ba73"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.545026 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x69lg" event={"ID":"0e8ff110-c94f-4dc0-8420-e00cff505a20","Type":"ContainerStarted","Data":"ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.546197 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83","Type":"ContainerStarted","Data":"4e8b5353de2704e3d9e2b5f98979ae26079fd9e26ed3dfabfee292a77115db72"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.547707 4726 generic.go:334] "Generic (PLEG): container finished" podID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerID="3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9" exitCode=0 Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.547780 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" event={"ID":"7f4503d8-cfb2-47db-8c7d-597566f57dfa","Type":"ContainerDied","Data":"3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.547805 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" event={"ID":"7f4503d8-cfb2-47db-8c7d-597566f57dfa","Type":"ContainerStarted","Data":"9492396cc852e7ede59792faae30074dff0a7fdfdb1e531a313918a73d53a596"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.549991 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m8rwl" event={"ID":"b3dba01d-46b8-407b-9622-5fd1534b9edc","Type":"ContainerStarted","Data":"491f5e8a8a85b7bc0611134218b495881366a5a89117ef627d978935edc8100e"} Oct 04 04:02:28 crc kubenswrapper[4726]: I1004 04:02:28.626957 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-m8rwl" podStartSLOduration=2.6269363869999998 podStartE2EDuration="2.626936387s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:28.591186021 +0000 UTC m=+1322.765809254" watchObservedRunningTime="2025-10-04 04:02:28.626936387 +0000 UTC m=+1322.801559600" Oct 04 04:02:29 crc kubenswrapper[4726]: I1004 04:02:29.566787 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x69lg" event={"ID":"0e8ff110-c94f-4dc0-8420-e00cff505a20","Type":"ContainerStarted","Data":"104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a"} Oct 04 04:02:29 crc kubenswrapper[4726]: I1004 04:02:29.573022 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" event={"ID":"7f4503d8-cfb2-47db-8c7d-597566f57dfa","Type":"ContainerStarted","Data":"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5"} Oct 04 04:02:29 crc kubenswrapper[4726]: I1004 04:02:29.573062 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:29 crc kubenswrapper[4726]: I1004 04:02:29.595601 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-x69lg" podStartSLOduration=2.595564292 podStartE2EDuration="2.595564292s" podCreationTimestamp="2025-10-04 04:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:29.594882043 +0000 UTC m=+1323.769505256" watchObservedRunningTime="2025-10-04 04:02:29.595564292 +0000 UTC m=+1323.770187505" Oct 04 04:02:29 crc kubenswrapper[4726]: I1004 04:02:29.620511 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" podStartSLOduration=3.620487769 podStartE2EDuration="3.620487769s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:29.614241447 +0000 UTC m=+1323.788864680" watchObservedRunningTime="2025-10-04 04:02:29.620487769 +0000 UTC m=+1323.795110982" Oct 04 04:02:30 crc kubenswrapper[4726]: I1004 04:02:30.372476 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:30 crc kubenswrapper[4726]: I1004 04:02:30.408852 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.594229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed51a4d9-f248-470f-a887-1ae137e39a66","Type":"ContainerStarted","Data":"d61e64a1a93a2ac275345ec3b2dc40dbce5172981bf6f137d340956567b48a3e"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.594301 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ed51a4d9-f248-470f-a887-1ae137e39a66" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d61e64a1a93a2ac275345ec3b2dc40dbce5172981bf6f137d340956567b48a3e" gracePeriod=30 Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.600469 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerStarted","Data":"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.600528 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerStarted","Data":"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.600574 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-metadata" containerID="cri-o://45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" gracePeriod=30 Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.600885 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-log" containerID="cri-o://4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" gracePeriod=30 Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.609948 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerStarted","Data":"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.610004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerStarted","Data":"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.620712 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83","Type":"ContainerStarted","Data":"4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082"} Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.628978 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.801741784 podStartE2EDuration="5.628956292s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="2025-10-04 04:02:27.590597206 +0000 UTC m=+1321.765220419" lastFinishedPulling="2025-10-04 04:02:30.417811714 +0000 UTC m=+1324.592434927" observedRunningTime="2025-10-04 04:02:31.616793047 +0000 UTC m=+1325.791416270" watchObservedRunningTime="2025-10-04 04:02:31.628956292 +0000 UTC m=+1325.803579525" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.666337 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.823967207 podStartE2EDuration="5.666317732s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="2025-10-04 04:02:27.598341699 +0000 UTC m=+1321.772964912" lastFinishedPulling="2025-10-04 04:02:30.440692224 +0000 UTC m=+1324.615315437" observedRunningTime="2025-10-04 04:02:31.643602476 +0000 UTC m=+1325.818225689" watchObservedRunningTime="2025-10-04 04:02:31.666317732 +0000 UTC m=+1325.840940945" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.673567 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.743826758 podStartE2EDuration="5.673551541s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="2025-10-04 04:02:27.487881785 +0000 UTC m=+1321.662505008" lastFinishedPulling="2025-10-04 04:02:30.417606578 +0000 UTC m=+1324.592229791" observedRunningTime="2025-10-04 04:02:31.662491876 +0000 UTC m=+1325.837115089" watchObservedRunningTime="2025-10-04 04:02:31.673551541 +0000 UTC m=+1325.848174754" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.689249 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.080169438 podStartE2EDuration="5.689233753s" podCreationTimestamp="2025-10-04 04:02:26 +0000 UTC" firstStartedPulling="2025-10-04 04:02:27.808871612 +0000 UTC m=+1321.983494825" lastFinishedPulling="2025-10-04 04:02:30.417935917 +0000 UTC m=+1324.592559140" observedRunningTime="2025-10-04 04:02:31.678400215 +0000 UTC m=+1325.853023438" watchObservedRunningTime="2025-10-04 04:02:31.689233753 +0000 UTC m=+1325.863856966" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.860216 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.860263 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:31 crc kubenswrapper[4726]: I1004 04:02:31.861288 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.173365 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.204325 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.405850 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs\") pod \"f1e768a1-9c89-43e7-8623-313afffc6a80\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.405953 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data\") pod \"f1e768a1-9c89-43e7-8623-313afffc6a80\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.406122 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle\") pod \"f1e768a1-9c89-43e7-8623-313afffc6a80\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.406178 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvzcg\" (UniqueName: \"kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg\") pod \"f1e768a1-9c89-43e7-8623-313afffc6a80\" (UID: \"f1e768a1-9c89-43e7-8623-313afffc6a80\") " Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.406226 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs" (OuterVolumeSpecName: "logs") pod "f1e768a1-9c89-43e7-8623-313afffc6a80" (UID: "f1e768a1-9c89-43e7-8623-313afffc6a80"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.406584 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e768a1-9c89-43e7-8623-313afffc6a80-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.411380 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg" (OuterVolumeSpecName: "kube-api-access-vvzcg") pod "f1e768a1-9c89-43e7-8623-313afffc6a80" (UID: "f1e768a1-9c89-43e7-8623-313afffc6a80"). InnerVolumeSpecName "kube-api-access-vvzcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.435496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1e768a1-9c89-43e7-8623-313afffc6a80" (UID: "f1e768a1-9c89-43e7-8623-313afffc6a80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.436212 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data" (OuterVolumeSpecName: "config-data") pod "f1e768a1-9c89-43e7-8623-313afffc6a80" (UID: "f1e768a1-9c89-43e7-8623-313afffc6a80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.509190 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.509227 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvzcg\" (UniqueName: \"kubernetes.io/projected/f1e768a1-9c89-43e7-8623-313afffc6a80-kube-api-access-vvzcg\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.509242 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e768a1-9c89-43e7-8623-313afffc6a80-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.570283 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.635742 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerID="45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" exitCode=0 Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.635775 4726 generic.go:334] "Generic (PLEG): container finished" podID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerID="4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" exitCode=143 Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.635828 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.637480 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerDied","Data":"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802"} Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.637519 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerDied","Data":"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053"} Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.637536 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1e768a1-9c89-43e7-8623-313afffc6a80","Type":"ContainerDied","Data":"ea0d45a51987715d2693de0f361553d3e99255e63a720fb97e309c375cf1ba73"} Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.637545 4726 scope.go:117] "RemoveContainer" containerID="45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.669924 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.671907 4726 scope.go:117] "RemoveContainer" containerID="4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.702262 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.710390 4726 scope.go:117] "RemoveContainer" containerID="45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" Oct 04 04:02:32 crc kubenswrapper[4726]: E1004 04:02:32.712345 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802\": container with ID starting with 45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802 not found: ID does not exist" containerID="45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.712388 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802"} err="failed to get container status \"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802\": rpc error: code = NotFound desc = could not find container \"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802\": container with ID starting with 45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802 not found: ID does not exist" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.712416 4726 scope.go:117] "RemoveContainer" containerID="4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" Oct 04 04:02:32 crc kubenswrapper[4726]: E1004 04:02:32.713064 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053\": container with ID starting with 4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053 not found: ID does not exist" containerID="4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.713095 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053"} err="failed to get container status \"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053\": rpc error: code = NotFound desc = could not find container \"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053\": container with ID starting with 4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053 not found: ID does not exist" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.713135 4726 scope.go:117] "RemoveContainer" containerID="45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.713593 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802"} err="failed to get container status \"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802\": rpc error: code = NotFound desc = could not find container \"45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802\": container with ID starting with 45aed0ef9e20aca26fbb95d24025f0cc9605fe759ef01ee6631a926a00689802 not found: ID does not exist" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.713639 4726 scope.go:117] "RemoveContainer" containerID="4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.713963 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053"} err="failed to get container status \"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053\": rpc error: code = NotFound desc = could not find container \"4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053\": container with ID starting with 4b8c516ed8d36a240ff3d85ede4b391411710f3e263b5d30f185d5514b098053 not found: ID does not exist" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.716228 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:32 crc kubenswrapper[4726]: E1004 04:02:32.716751 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-metadata" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.716779 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-metadata" Oct 04 04:02:32 crc kubenswrapper[4726]: E1004 04:02:32.716809 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-log" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.716819 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-log" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.717056 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-log" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.717097 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" containerName="nova-metadata-metadata" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.718445 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.722991 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.723253 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.728300 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.827075 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.827148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.827179 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.827232 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.828324 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml2jr\" (UniqueName: \"kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.930180 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml2jr\" (UniqueName: \"kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.930676 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.930709 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.930749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.930817 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.932043 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.934756 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.936693 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.936918 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:32 crc kubenswrapper[4726]: I1004 04:02:32.947147 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml2jr\" (UniqueName: \"kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr\") pod \"nova-metadata-0\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " pod="openstack/nova-metadata-0" Oct 04 04:02:33 crc kubenswrapper[4726]: I1004 04:02:33.052248 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:33 crc kubenswrapper[4726]: I1004 04:02:33.526644 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:33 crc kubenswrapper[4726]: W1004 04:02:33.529261 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65f5b6aa_4620_4067_9711_f1fca3587c30.slice/crio-2c67dec905a2a6e91b40f8d735cf6c42889511c95fee9d44699b3ad9e8c4fc4b WatchSource:0}: Error finding container 2c67dec905a2a6e91b40f8d735cf6c42889511c95fee9d44699b3ad9e8c4fc4b: Status 404 returned error can't find the container with id 2c67dec905a2a6e91b40f8d735cf6c42889511c95fee9d44699b3ad9e8c4fc4b Oct 04 04:02:33 crc kubenswrapper[4726]: I1004 04:02:33.652211 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerStarted","Data":"2c67dec905a2a6e91b40f8d735cf6c42889511c95fee9d44699b3ad9e8c4fc4b"} Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.187981 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.188306 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.524277 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e768a1-9c89-43e7-8623-313afffc6a80" path="/var/lib/kubelet/pods/f1e768a1-9c89-43e7-8623-313afffc6a80/volumes" Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.675466 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerStarted","Data":"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889"} Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.676494 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerStarted","Data":"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c"} Oct 04 04:02:34 crc kubenswrapper[4726]: I1004 04:02:34.704182 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.704157655 podStartE2EDuration="2.704157655s" podCreationTimestamp="2025-10-04 04:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:34.696950697 +0000 UTC m=+1328.871573940" watchObservedRunningTime="2025-10-04 04:02:34.704157655 +0000 UTC m=+1328.878780878" Oct 04 04:02:35 crc kubenswrapper[4726]: I1004 04:02:35.688135 4726 generic.go:334] "Generic (PLEG): container finished" podID="b3dba01d-46b8-407b-9622-5fd1534b9edc" containerID="491f5e8a8a85b7bc0611134218b495881366a5a89117ef627d978935edc8100e" exitCode=0 Oct 04 04:02:35 crc kubenswrapper[4726]: I1004 04:02:35.688212 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m8rwl" event={"ID":"b3dba01d-46b8-407b-9622-5fd1534b9edc","Type":"ContainerDied","Data":"491f5e8a8a85b7bc0611134218b495881366a5a89117ef627d978935edc8100e"} Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.416883 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.417223 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" containerName="kube-state-metrics" containerID="cri-o://c103577951bb41470de5fc25898bdba5b2c1c455d78979280fbdddf0ca82d737" gracePeriod=30 Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.704941 4726 generic.go:334] "Generic (PLEG): container finished" podID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" containerID="c103577951bb41470de5fc25898bdba5b2c1c455d78979280fbdddf0ca82d737" exitCode=2 Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.705013 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5d1c61fb-2030-40d7-a859-7f8c30983fc1","Type":"ContainerDied","Data":"c103577951bb41470de5fc25898bdba5b2c1c455d78979280fbdddf0ca82d737"} Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.850469 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:02:36 crc kubenswrapper[4726]: I1004 04:02:36.850743 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.054965 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.059950 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.170273 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.172586 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.208954 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.211438 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts\") pod \"b3dba01d-46b8-407b-9622-5fd1534b9edc\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.211505 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgb5g\" (UniqueName: \"kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g\") pod \"5d1c61fb-2030-40d7-a859-7f8c30983fc1\" (UID: \"5d1c61fb-2030-40d7-a859-7f8c30983fc1\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.211586 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle\") pod \"b3dba01d-46b8-407b-9622-5fd1534b9edc\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.211619 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt6m4\" (UniqueName: \"kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4\") pod \"b3dba01d-46b8-407b-9622-5fd1534b9edc\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.211715 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data\") pod \"b3dba01d-46b8-407b-9622-5fd1534b9edc\" (UID: \"b3dba01d-46b8-407b-9622-5fd1534b9edc\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.221215 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts" (OuterVolumeSpecName: "scripts") pod "b3dba01d-46b8-407b-9622-5fd1534b9edc" (UID: "b3dba01d-46b8-407b-9622-5fd1534b9edc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.226423 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.226846 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="dnsmasq-dns" containerID="cri-o://36c16ca90803489bd416b0a75f9350c121f6743ad6e3bc03208b982406944302" gracePeriod=10 Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.239867 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g" (OuterVolumeSpecName: "kube-api-access-vgb5g") pod "5d1c61fb-2030-40d7-a859-7f8c30983fc1" (UID: "5d1c61fb-2030-40d7-a859-7f8c30983fc1"). InnerVolumeSpecName "kube-api-access-vgb5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.251832 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4" (OuterVolumeSpecName: "kube-api-access-tt6m4") pod "b3dba01d-46b8-407b-9622-5fd1534b9edc" (UID: "b3dba01d-46b8-407b-9622-5fd1534b9edc"). InnerVolumeSpecName "kube-api-access-tt6m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.280761 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data" (OuterVolumeSpecName: "config-data") pod "b3dba01d-46b8-407b-9622-5fd1534b9edc" (UID: "b3dba01d-46b8-407b-9622-5fd1534b9edc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.280951 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3dba01d-46b8-407b-9622-5fd1534b9edc" (UID: "b3dba01d-46b8-407b-9622-5fd1534b9edc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.312992 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.313019 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.313029 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgb5g\" (UniqueName: \"kubernetes.io/projected/5d1c61fb-2030-40d7-a859-7f8c30983fc1-kube-api-access-vgb5g\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.313038 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3dba01d-46b8-407b-9622-5fd1534b9edc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.313048 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt6m4\" (UniqueName: \"kubernetes.io/projected/b3dba01d-46b8-407b-9622-5fd1534b9edc-kube-api-access-tt6m4\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.714052 4726 generic.go:334] "Generic (PLEG): container finished" podID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerID="36c16ca90803489bd416b0a75f9350c121f6743ad6e3bc03208b982406944302" exitCode=0 Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.714148 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" event={"ID":"5f5f1547-f888-4cc7-97f3-84ce9117f5c8","Type":"ContainerDied","Data":"36c16ca90803489bd416b0a75f9350c121f6743ad6e3bc03208b982406944302"} Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.715043 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" event={"ID":"5f5f1547-f888-4cc7-97f3-84ce9117f5c8","Type":"ContainerDied","Data":"fa8fa0908805788706aa6427e09c2fe9525b3d1d0946b3b128fec06d9336efdb"} Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.715150 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa8fa0908805788706aa6427e09c2fe9525b3d1d0946b3b128fec06d9336efdb" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.716586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"5d1c61fb-2030-40d7-a859-7f8c30983fc1","Type":"ContainerDied","Data":"a5d8ed0bb822cb8fae7ec8a90976a8f6408d1451aa2d3c2ae179d2bc7f5de0d9"} Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.716601 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.716650 4726 scope.go:117] "RemoveContainer" containerID="c103577951bb41470de5fc25898bdba5b2c1c455d78979280fbdddf0ca82d737" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.718739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m8rwl" event={"ID":"b3dba01d-46b8-407b-9622-5fd1534b9edc","Type":"ContainerDied","Data":"31637f9dd1d67eaf73e05aa3ece1f84f39e8adc4904f1418b812ea0e9c05fd95"} Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.718842 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31637f9dd1d67eaf73e05aa3ece1f84f39e8adc4904f1418b812ea0e9c05fd95" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.718906 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m8rwl" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.762634 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.834887 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.848863 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.857871 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858513 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858600 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858635 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858688 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858732 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfk9j\" (UniqueName: \"kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.858772 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0\") pod \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\" (UID: \"5f5f1547-f888-4cc7-97f3-84ce9117f5c8\") " Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.913467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j" (OuterVolumeSpecName: "kube-api-access-nfk9j") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "kube-api-access-nfk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.913528 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:37 crc kubenswrapper[4726]: E1004 04:02:37.913977 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="init" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.913991 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="init" Oct 04 04:02:37 crc kubenswrapper[4726]: E1004 04:02:37.914004 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3dba01d-46b8-407b-9622-5fd1534b9edc" containerName="nova-manage" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914010 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3dba01d-46b8-407b-9622-5fd1534b9edc" containerName="nova-manage" Oct 04 04:02:37 crc kubenswrapper[4726]: E1004 04:02:37.914044 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="dnsmasq-dns" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914050 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="dnsmasq-dns" Oct 04 04:02:37 crc kubenswrapper[4726]: E1004 04:02:37.914064 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" containerName="kube-state-metrics" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914070 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" containerName="kube-state-metrics" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914260 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" containerName="dnsmasq-dns" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914289 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" containerName="kube-state-metrics" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914299 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3dba01d-46b8-407b-9622-5fd1534b9edc" containerName="nova-manage" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.914885 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.918055 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.918345 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.941420 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.941700 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.958369 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.960865 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.960911 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.960984 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.961045 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp522\" (UniqueName: \"kubernetes.io/projected/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-api-access-tp522\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.961123 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfk9j\" (UniqueName: \"kubernetes.io/projected/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-kube-api-access-nfk9j\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.995371 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:02:37 crc kubenswrapper[4726]: I1004 04:02:37.996820 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.047612 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config" (OuterVolumeSpecName: "config") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.049013 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.054369 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.055561 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.059170 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.064676 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5f5f1547-f888-4cc7-97f3-84ce9117f5c8" (UID: "5f5f1547-f888-4cc7-97f3-84ce9117f5c8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072071 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072204 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp522\" (UniqueName: \"kubernetes.io/projected/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-api-access-tp522\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072300 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072388 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072399 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072408 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072418 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.072425 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f5f1547-f888-4cc7-97f3-84ce9117f5c8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.088842 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.088985 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.093868 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.094136 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.109997 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp522\" (UniqueName: \"kubernetes.io/projected/c4db3f45-a9c5-4092-a236-f3c345e2303d-kube-api-access-tp522\") pod \"kube-state-metrics-0\" (UID: \"c4db3f45-a9c5-4092-a236-f3c345e2303d\") " pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.311168 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:02:38 crc kubenswrapper[4726]: E1004 04:02:38.338483 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e8ff110_c94f_4dc0_8420_e00cff505a20.slice/crio-104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e8ff110_c94f_4dc0_8420_e00cff505a20.slice/crio-conmon-104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.433878 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.519493 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d1c61fb-2030-40d7-a859-7f8c30983fc1" path="/var/lib/kubelet/pods/5d1c61fb-2030-40d7-a859-7f8c30983fc1/volumes" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.716616 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.716939 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-central-agent" containerID="cri-o://f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.717003 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="proxy-httpd" containerID="cri-o://cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.717018 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="sg-core" containerID="cri-o://89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.717055 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-notification-agent" containerID="cri-o://d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.728868 4726 generic.go:334] "Generic (PLEG): container finished" podID="0e8ff110-c94f-4dc0-8420-e00cff505a20" containerID="104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a" exitCode=0 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.728948 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x69lg" event={"ID":"0e8ff110-c94f-4dc0-8420-e00cff505a20","Type":"ContainerDied","Data":"104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a"} Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.731097 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-mh5qd" Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.731266 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-log" containerID="cri-o://c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.731334 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-api" containerID="cri-o://77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6" gracePeriod=30 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.770730 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:02:38 crc kubenswrapper[4726]: W1004 04:02:38.802374 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4db3f45_a9c5_4092_a236_f3c345e2303d.slice/crio-78fb13f8b809816491407fc3361cf1d53f9a87193069a980355a9c75e15cb607 WatchSource:0}: Error finding container 78fb13f8b809816491407fc3361cf1d53f9a87193069a980355a9c75e15cb607: Status 404 returned error can't find the container with id 78fb13f8b809816491407fc3361cf1d53f9a87193069a980355a9c75e15cb607 Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.818546 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-mh5qd"] Oct 04 04:02:38 crc kubenswrapper[4726]: I1004 04:02:38.823735 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741352 4726 generic.go:334] "Generic (PLEG): container finished" podID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerID="cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63" exitCode=0 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741681 4726 generic.go:334] "Generic (PLEG): container finished" podID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerID="89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d" exitCode=2 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741696 4726 generic.go:334] "Generic (PLEG): container finished" podID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerID="f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914" exitCode=0 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741537 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerDied","Data":"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741775 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerDied","Data":"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.741795 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerDied","Data":"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.743565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c4db3f45-a9c5-4092-a236-f3c345e2303d","Type":"ContainerStarted","Data":"3d1286ed80bca77cdd2395d9e617a1a4d8c82d24e31490f2fdf218f966821bce"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.743604 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c4db3f45-a9c5-4092-a236-f3c345e2303d","Type":"ContainerStarted","Data":"78fb13f8b809816491407fc3361cf1d53f9a87193069a980355a9c75e15cb607"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.744843 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.748373 4726 generic.go:334] "Generic (PLEG): container finished" podID="f55722bd-a894-4195-b302-1a56c45b8579" containerID="c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f" exitCode=143 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.748565 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerDied","Data":"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f"} Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.748696 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerName="nova-scheduler-scheduler" containerID="cri-o://4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" gracePeriod=30 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.749014 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-log" containerID="cri-o://ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" gracePeriod=30 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.749147 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-metadata" containerID="cri-o://9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" gracePeriod=30 Oct 04 04:02:39 crc kubenswrapper[4726]: I1004 04:02:39.762806 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.352541574 podStartE2EDuration="2.762792821s" podCreationTimestamp="2025-10-04 04:02:37 +0000 UTC" firstStartedPulling="2025-10-04 04:02:38.805136007 +0000 UTC m=+1332.979759210" lastFinishedPulling="2025-10-04 04:02:39.215387244 +0000 UTC m=+1333.390010457" observedRunningTime="2025-10-04 04:02:39.760874558 +0000 UTC m=+1333.935497761" watchObservedRunningTime="2025-10-04 04:02:39.762792821 +0000 UTC m=+1333.937416034" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.130085 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.293653 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.323555 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfhmv\" (UniqueName: \"kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv\") pod \"0e8ff110-c94f-4dc0-8420-e00cff505a20\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.323766 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data\") pod \"0e8ff110-c94f-4dc0-8420-e00cff505a20\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.323873 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts\") pod \"0e8ff110-c94f-4dc0-8420-e00cff505a20\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.323890 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle\") pod \"0e8ff110-c94f-4dc0-8420-e00cff505a20\" (UID: \"0e8ff110-c94f-4dc0-8420-e00cff505a20\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.329225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv" (OuterVolumeSpecName: "kube-api-access-rfhmv") pod "0e8ff110-c94f-4dc0-8420-e00cff505a20" (UID: "0e8ff110-c94f-4dc0-8420-e00cff505a20"). InnerVolumeSpecName "kube-api-access-rfhmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.329589 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts" (OuterVolumeSpecName: "scripts") pod "0e8ff110-c94f-4dc0-8420-e00cff505a20" (UID: "0e8ff110-c94f-4dc0-8420-e00cff505a20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.349535 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e8ff110-c94f-4dc0-8420-e00cff505a20" (UID: "0e8ff110-c94f-4dc0-8420-e00cff505a20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.351394 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data" (OuterVolumeSpecName: "config-data") pod "0e8ff110-c94f-4dc0-8420-e00cff505a20" (UID: "0e8ff110-c94f-4dc0-8420-e00cff505a20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425061 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs\") pod \"65f5b6aa-4620-4067-9711-f1fca3587c30\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425186 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml2jr\" (UniqueName: \"kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr\") pod \"65f5b6aa-4620-4067-9711-f1fca3587c30\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425222 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle\") pod \"65f5b6aa-4620-4067-9711-f1fca3587c30\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs\") pod \"65f5b6aa-4620-4067-9711-f1fca3587c30\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425350 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data\") pod \"65f5b6aa-4620-4067-9711-f1fca3587c30\" (UID: \"65f5b6aa-4620-4067-9711-f1fca3587c30\") " Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425806 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425823 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425834 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e8ff110-c94f-4dc0-8420-e00cff505a20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.425848 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfhmv\" (UniqueName: \"kubernetes.io/projected/0e8ff110-c94f-4dc0-8420-e00cff505a20-kube-api-access-rfhmv\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.426120 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs" (OuterVolumeSpecName: "logs") pod "65f5b6aa-4620-4067-9711-f1fca3587c30" (UID: "65f5b6aa-4620-4067-9711-f1fca3587c30"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.429693 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr" (OuterVolumeSpecName: "kube-api-access-ml2jr") pod "65f5b6aa-4620-4067-9711-f1fca3587c30" (UID: "65f5b6aa-4620-4067-9711-f1fca3587c30"). InnerVolumeSpecName "kube-api-access-ml2jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.449252 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65f5b6aa-4620-4067-9711-f1fca3587c30" (UID: "65f5b6aa-4620-4067-9711-f1fca3587c30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.450299 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data" (OuterVolumeSpecName: "config-data") pod "65f5b6aa-4620-4067-9711-f1fca3587c30" (UID: "65f5b6aa-4620-4067-9711-f1fca3587c30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.486678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "65f5b6aa-4620-4067-9711-f1fca3587c30" (UID: "65f5b6aa-4620-4067-9711-f1fca3587c30"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.513139 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f5f1547-f888-4cc7-97f3-84ce9117f5c8" path="/var/lib/kubelet/pods/5f5f1547-f888-4cc7-97f3-84ce9117f5c8/volumes" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.527027 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml2jr\" (UniqueName: \"kubernetes.io/projected/65f5b6aa-4620-4067-9711-f1fca3587c30-kube-api-access-ml2jr\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.527064 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.527074 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.527083 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65f5b6aa-4620-4067-9711-f1fca3587c30-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.527094 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65f5b6aa-4620-4067-9711-f1fca3587c30-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760338 4726 generic.go:334] "Generic (PLEG): container finished" podID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerID="9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" exitCode=0 Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760372 4726 generic.go:334] "Generic (PLEG): container finished" podID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerID="ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" exitCode=143 Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760382 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760399 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerDied","Data":"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889"} Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760474 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerDied","Data":"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c"} Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760489 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"65f5b6aa-4620-4067-9711-f1fca3587c30","Type":"ContainerDied","Data":"2c67dec905a2a6e91b40f8d735cf6c42889511c95fee9d44699b3ad9e8c4fc4b"} Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.760509 4726 scope.go:117] "RemoveContainer" containerID="9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.763221 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x69lg" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.763461 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x69lg" event={"ID":"0e8ff110-c94f-4dc0-8420-e00cff505a20","Type":"ContainerDied","Data":"ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55"} Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.763564 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec77b91613c47aa1aadbe778f067537f036493acb4615ec528aa6d23cd447b55" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.791941 4726 scope.go:117] "RemoveContainer" containerID="ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.812921 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.833019 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.843547 4726 scope.go:117] "RemoveContainer" containerID="9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" Oct 04 04:02:40 crc kubenswrapper[4726]: E1004 04:02:40.843978 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889\": container with ID starting with 9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889 not found: ID does not exist" containerID="9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844026 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889"} err="failed to get container status \"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889\": rpc error: code = NotFound desc = could not find container \"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889\": container with ID starting with 9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889 not found: ID does not exist" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844056 4726 scope.go:117] "RemoveContainer" containerID="ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" Oct 04 04:02:40 crc kubenswrapper[4726]: E1004 04:02:40.844297 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c\": container with ID starting with ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c not found: ID does not exist" containerID="ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844319 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c"} err="failed to get container status \"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c\": rpc error: code = NotFound desc = could not find container \"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c\": container with ID starting with ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c not found: ID does not exist" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844332 4726 scope.go:117] "RemoveContainer" containerID="9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844595 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889"} err="failed to get container status \"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889\": rpc error: code = NotFound desc = could not find container \"9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889\": container with ID starting with 9a7e55f6c06eb5303563e4699c755322ee90b26ea477a2dc3b82f2df7e616889 not found: ID does not exist" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844628 4726 scope.go:117] "RemoveContainer" containerID="ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.844861 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c"} err="failed to get container status \"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c\": rpc error: code = NotFound desc = could not find container \"ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c\": container with ID starting with ad3d8efea091bcf3eb522f75f9429b1f0907be0d862286a4a847c944499d1b9c not found: ID does not exist" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.849875 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: E1004 04:02:40.850368 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e8ff110-c94f-4dc0-8420-e00cff505a20" containerName="nova-cell1-conductor-db-sync" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850388 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e8ff110-c94f-4dc0-8420-e00cff505a20" containerName="nova-cell1-conductor-db-sync" Oct 04 04:02:40 crc kubenswrapper[4726]: E1004 04:02:40.850413 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-log" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850422 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-log" Oct 04 04:02:40 crc kubenswrapper[4726]: E1004 04:02:40.850449 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-metadata" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850458 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-metadata" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850707 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e8ff110-c94f-4dc0-8420-e00cff505a20" containerName="nova-cell1-conductor-db-sync" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850728 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-metadata" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.850749 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" containerName="nova-metadata-log" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.851963 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.855432 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.855645 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.863265 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.870234 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.870315 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.870396 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m227\" (UniqueName: \"kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.870427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.870458 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.872270 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.874264 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.878820 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.885804 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972062 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972141 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvrjk\" (UniqueName: \"kubernetes.io/projected/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-kube-api-access-nvrjk\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972164 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972185 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972202 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972231 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m227\" (UniqueName: \"kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972249 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.972271 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.973284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.977925 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.979879 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.981290 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:40 crc kubenswrapper[4726]: I1004 04:02:40.994055 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m227\" (UniqueName: \"kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227\") pod \"nova-metadata-0\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " pod="openstack/nova-metadata-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.079171 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvrjk\" (UniqueName: \"kubernetes.io/projected/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-kube-api-access-nvrjk\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.079233 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.079262 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.091998 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.100793 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.100866 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvrjk\" (UniqueName: \"kubernetes.io/projected/11f9ad5d-1851-40d1-b8f9-cfd476b498ce-kube-api-access-nvrjk\") pod \"nova-cell1-conductor-0\" (UID: \"11f9ad5d-1851-40d1-b8f9-cfd476b498ce\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.199553 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.203564 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.209176 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281338 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281461 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281516 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zns99\" (UniqueName: \"kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281582 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281649 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281682 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281691 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.281819 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts\") pod \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\" (UID: \"db2708ed-c9c1-4d6d-83ab-d48c621ed688\") " Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.282202 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.282665 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.282694 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db2708ed-c9c1-4d6d-83ab-d48c621ed688-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.295361 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts" (OuterVolumeSpecName: "scripts") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.295443 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99" (OuterVolumeSpecName: "kube-api-access-zns99") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "kube-api-access-zns99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.326418 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.372935 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.385659 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zns99\" (UniqueName: \"kubernetes.io/projected/db2708ed-c9c1-4d6d-83ab-d48c621ed688-kube-api-access-zns99\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.385691 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.385700 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.385710 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.413145 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data" (OuterVolumeSpecName: "config-data") pod "db2708ed-c9c1-4d6d-83ab-d48c621ed688" (UID: "db2708ed-c9c1-4d6d-83ab-d48c621ed688"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.487826 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db2708ed-c9c1-4d6d-83ab-d48c621ed688-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.691289 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.760890 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.779251 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"11f9ad5d-1851-40d1-b8f9-cfd476b498ce","Type":"ContainerStarted","Data":"76b2c87a92c51508877e700d70ea3e4ad8728f3edddd32c31365a8e3d55bc137"} Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.784595 4726 generic.go:334] "Generic (PLEG): container finished" podID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerID="d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e" exitCode=0 Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.785608 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerDied","Data":"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e"} Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.785681 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db2708ed-c9c1-4d6d-83ab-d48c621ed688","Type":"ContainerDied","Data":"563467c6512879817f649766b69c8ce7b5bd4e8f304b05cdd1c1c26252647877"} Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.785691 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.785722 4726 scope.go:117] "RemoveContainer" containerID="cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.882771 4726 scope.go:117] "RemoveContainer" containerID="89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d" Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.980865 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.989345 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:41 crc kubenswrapper[4726]: I1004 04:02:41.997185 4726 scope.go:117] "RemoveContainer" containerID="d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.004537 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.005174 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-notification-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005189 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-notification-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.005198 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="proxy-httpd" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005205 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="proxy-httpd" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.005213 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-central-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005219 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-central-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.005231 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="sg-core" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005236 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="sg-core" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005398 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-central-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005416 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="ceilometer-notification-agent" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005436 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="proxy-httpd" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.005443 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" containerName="sg-core" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.007379 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.010502 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.010795 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.010922 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.012691 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.038311 4726 scope.go:117] "RemoveContainer" containerID="f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.061973 4726 scope.go:117] "RemoveContainer" containerID="cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.062315 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63\": container with ID starting with cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63 not found: ID does not exist" containerID="cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062347 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63"} err="failed to get container status \"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63\": rpc error: code = NotFound desc = could not find container \"cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63\": container with ID starting with cc3dd2502b15bd218ef770419837b65d659b1661283005c3c290deef78974b63 not found: ID does not exist" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062369 4726 scope.go:117] "RemoveContainer" containerID="89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.062712 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d\": container with ID starting with 89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d not found: ID does not exist" containerID="89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062738 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d"} err="failed to get container status \"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d\": rpc error: code = NotFound desc = could not find container \"89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d\": container with ID starting with 89034ce483aa7aa4eb314577978c98b18a77641ebf69f94a066f61ea98f50c5d not found: ID does not exist" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062753 4726 scope.go:117] "RemoveContainer" containerID="d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.062939 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e\": container with ID starting with d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e not found: ID does not exist" containerID="d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062967 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e"} err="failed to get container status \"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e\": rpc error: code = NotFound desc = could not find container \"d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e\": container with ID starting with d3ee456db9fb8d7a8b1a7cb6ccb8d780a5c16fa9c350f7aac0025e107c7b9a5e not found: ID does not exist" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.062981 4726 scope.go:117] "RemoveContainer" containerID="f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.063558 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914\": container with ID starting with f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914 not found: ID does not exist" containerID="f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.063581 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914"} err="failed to get container status \"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914\": rpc error: code = NotFound desc = could not find container \"f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914\": container with ID starting with f108d589360487f5c37b8d591bef7a62deca02869b14022368e434f7dab6b914 not found: ID does not exist" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098275 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098339 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098371 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098391 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86nmn\" (UniqueName: \"kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098585 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098699 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098764 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.098955 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.174822 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.178486 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.180553 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:02:42 crc kubenswrapper[4726]: E1004 04:02:42.180589 4726 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerName="nova-scheduler-scheduler" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.200680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.200766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.200841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.201362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.201426 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.201447 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86nmn\" (UniqueName: \"kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.201927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.201981 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.202013 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.202294 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.205263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.206194 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.206660 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.208390 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.209573 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.218698 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86nmn\" (UniqueName: \"kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn\") pod \"ceilometer-0\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.330679 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.528142 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f5b6aa-4620-4067-9711-f1fca3587c30" path="/var/lib/kubelet/pods/65f5b6aa-4620-4067-9711-f1fca3587c30/volumes" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.529930 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db2708ed-c9c1-4d6d-83ab-d48c621ed688" path="/var/lib/kubelet/pods/db2708ed-c9c1-4d6d-83ab-d48c621ed688/volumes" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.802991 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.814008 4726 generic.go:334] "Generic (PLEG): container finished" podID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerID="4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" exitCode=0 Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.814062 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83","Type":"ContainerDied","Data":"4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082"} Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.816064 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerStarted","Data":"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156"} Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.816086 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerStarted","Data":"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a"} Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.816096 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerStarted","Data":"12330fae995ea4f279ad2b25c805fd81984dc49ea80394ea6d917bc471fdfced"} Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.819534 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"11f9ad5d-1851-40d1-b8f9-cfd476b498ce","Type":"ContainerStarted","Data":"eff5854151201eb2896980c091607281a2fcbeda7d640f498826d2e686426791"} Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.820510 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.847800 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.847777023 podStartE2EDuration="2.847777023s" podCreationTimestamp="2025-10-04 04:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:42.837616102 +0000 UTC m=+1337.012239325" watchObservedRunningTime="2025-10-04 04:02:42.847777023 +0000 UTC m=+1337.022400246" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.861877 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.86185242 podStartE2EDuration="2.86185242s" podCreationTimestamp="2025-10-04 04:02:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:42.850870698 +0000 UTC m=+1337.025493921" watchObservedRunningTime="2025-10-04 04:02:42.86185242 +0000 UTC m=+1337.036475643" Oct 04 04:02:42 crc kubenswrapper[4726]: I1004 04:02:42.945221 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.017254 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data\") pod \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.017297 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4gsd\" (UniqueName: \"kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd\") pod \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.017357 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle\") pod \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\" (UID: \"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.023725 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd" (OuterVolumeSpecName: "kube-api-access-b4gsd") pod "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" (UID: "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83"). InnerVolumeSpecName "kube-api-access-b4gsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.048257 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data" (OuterVolumeSpecName: "config-data") pod "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" (UID: "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.055758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" (UID: "8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.119332 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.119373 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4gsd\" (UniqueName: \"kubernetes.io/projected/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-kube-api-access-b4gsd\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.119385 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.532687 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.628504 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs\") pod \"f55722bd-a894-4195-b302-1a56c45b8579\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.628585 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data\") pod \"f55722bd-a894-4195-b302-1a56c45b8579\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.628625 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcs2f\" (UniqueName: \"kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f\") pod \"f55722bd-a894-4195-b302-1a56c45b8579\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.628655 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle\") pod \"f55722bd-a894-4195-b302-1a56c45b8579\" (UID: \"f55722bd-a894-4195-b302-1a56c45b8579\") " Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.628973 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs" (OuterVolumeSpecName: "logs") pod "f55722bd-a894-4195-b302-1a56c45b8579" (UID: "f55722bd-a894-4195-b302-1a56c45b8579"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.629756 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f55722bd-a894-4195-b302-1a56c45b8579-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.636366 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f" (OuterVolumeSpecName: "kube-api-access-kcs2f") pod "f55722bd-a894-4195-b302-1a56c45b8579" (UID: "f55722bd-a894-4195-b302-1a56c45b8579"). InnerVolumeSpecName "kube-api-access-kcs2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.657180 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data" (OuterVolumeSpecName: "config-data") pod "f55722bd-a894-4195-b302-1a56c45b8579" (UID: "f55722bd-a894-4195-b302-1a56c45b8579"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.667209 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f55722bd-a894-4195-b302-1a56c45b8579" (UID: "f55722bd-a894-4195-b302-1a56c45b8579"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.731650 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.731682 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcs2f\" (UniqueName: \"kubernetes.io/projected/f55722bd-a894-4195-b302-1a56c45b8579-kube-api-access-kcs2f\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.731692 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55722bd-a894-4195-b302-1a56c45b8579-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.828864 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerStarted","Data":"f253fc1f2e654ca1867a5d40e2751f7f2418732264d2b0d727a511e87d4a23a6"} Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.830522 4726 generic.go:334] "Generic (PLEG): container finished" podID="f55722bd-a894-4195-b302-1a56c45b8579" containerID="77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6" exitCode=0 Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.830588 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerDied","Data":"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6"} Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.830615 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f55722bd-a894-4195-b302-1a56c45b8579","Type":"ContainerDied","Data":"f537c14de8809f86bd412ddb10d24b00787cb51f9e801d4ead99fc31dd9c2e27"} Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.830632 4726 scope.go:117] "RemoveContainer" containerID="77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.830730 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.836321 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.836915 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83","Type":"ContainerDied","Data":"4e8b5353de2704e3d9e2b5f98979ae26079fd9e26ed3dfabfee292a77115db72"} Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.858875 4726 scope.go:117] "RemoveContainer" containerID="c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.863874 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.875237 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.882323 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.886931 4726 scope.go:117] "RemoveContainer" containerID="77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6" Oct 04 04:02:43 crc kubenswrapper[4726]: E1004 04:02:43.887382 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6\": container with ID starting with 77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6 not found: ID does not exist" containerID="77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.887409 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6"} err="failed to get container status \"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6\": rpc error: code = NotFound desc = could not find container \"77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6\": container with ID starting with 77f3a6057d410d887c47c300c2a0278ec00f37aab0cc91714c41d4bc1f42ece6 not found: ID does not exist" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.887429 4726 scope.go:117] "RemoveContainer" containerID="c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f" Oct 04 04:02:43 crc kubenswrapper[4726]: E1004 04:02:43.887657 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f\": container with ID starting with c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f not found: ID does not exist" containerID="c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.887680 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f"} err="failed to get container status \"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f\": rpc error: code = NotFound desc = could not find container \"c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f\": container with ID starting with c9d867b202655e537e73c211fc8e4ea2e5aa4fe1d87af1cee97c6403a7d2c98f not found: ID does not exist" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.887692 4726 scope.go:117] "RemoveContainer" containerID="4dc4952e39e89a331d273a65dc43865bba6305e3c84701f4e4a017efaf17b082" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.898317 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918153 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: E1004 04:02:43.918542 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-api" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918558 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-api" Oct 04 04:02:43 crc kubenswrapper[4726]: E1004 04:02:43.918573 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerName="nova-scheduler-scheduler" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918582 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerName="nova-scheduler-scheduler" Oct 04 04:02:43 crc kubenswrapper[4726]: E1004 04:02:43.918593 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-log" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918599 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-log" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918766 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-api" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918785 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55722bd-a894-4195-b302-1a56c45b8579" containerName="nova-api-log" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.918806 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" containerName="nova-scheduler-scheduler" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.919711 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.925811 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.926326 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.927550 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.933474 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.978221 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:43 crc kubenswrapper[4726]: I1004 04:02:43.999301 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.049790 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hddbb\" (UniqueName: \"kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.050183 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.050407 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.050607 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.050782 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.051192 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.051810 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qwkl\" (UniqueName: \"kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.154342 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qwkl\" (UniqueName: \"kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.154867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hddbb\" (UniqueName: \"kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.155050 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.155300 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.155493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.155680 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.155892 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.156041 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.159313 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.159784 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.160383 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.160551 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.182536 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qwkl\" (UniqueName: \"kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl\") pod \"nova-scheduler-0\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.189490 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hddbb\" (UniqueName: \"kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb\") pod \"nova-api-0\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.270967 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.277895 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.520086 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83" path="/var/lib/kubelet/pods/8a70d7d0-c310-4e80-ab99-4e1b7f1f2b83/volumes" Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.522182 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f55722bd-a894-4195-b302-1a56c45b8579" path="/var/lib/kubelet/pods/f55722bd-a894-4195-b302-1a56c45b8579/volumes" Oct 04 04:02:44 crc kubenswrapper[4726]: W1004 04:02:44.782301 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode20838cb_bc7d_4917_b2fa_e2a1825b3ba3.slice/crio-a341266fdc9cd9ee5731ff987226d13947986a107a22d2c6a2bd9f315d3261a3 WatchSource:0}: Error finding container a341266fdc9cd9ee5731ff987226d13947986a107a22d2c6a2bd9f315d3261a3: Status 404 returned error can't find the container with id a341266fdc9cd9ee5731ff987226d13947986a107a22d2c6a2bd9f315d3261a3 Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.790462 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:02:44 crc kubenswrapper[4726]: W1004 04:02:44.803374 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4fb1f7b_e26c_4f9d_b3aa_efa89e23c03f.slice/crio-70258491387eb9fc9854dd47b76f47daab326706d6a47d5a0a8d9c283ed0e50a WatchSource:0}: Error finding container 70258491387eb9fc9854dd47b76f47daab326706d6a47d5a0a8d9c283ed0e50a: Status 404 returned error can't find the container with id 70258491387eb9fc9854dd47b76f47daab326706d6a47d5a0a8d9c283ed0e50a Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.809689 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.853562 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerStarted","Data":"ca6c6ce3b9f0ecfe725e6a57a4de9eb79e470a0e269311612c1cc4afeabb98dd"} Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.854772 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerStarted","Data":"70258491387eb9fc9854dd47b76f47daab326706d6a47d5a0a8d9c283ed0e50a"} Oct 04 04:02:44 crc kubenswrapper[4726]: I1004 04:02:44.858577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3","Type":"ContainerStarted","Data":"a341266fdc9cd9ee5731ff987226d13947986a107a22d2c6a2bd9f315d3261a3"} Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.869242 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerStarted","Data":"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4"} Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.870934 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerStarted","Data":"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca"} Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.872929 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3","Type":"ContainerStarted","Data":"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b"} Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.875891 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerStarted","Data":"88ce55993e6aaf7ebed96ae8b671f5efa5f400e6ddc564ee5db20bb74ea7f124"} Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.894221 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.894201372 podStartE2EDuration="2.894201372s" podCreationTimestamp="2025-10-04 04:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:45.892974398 +0000 UTC m=+1340.067597651" watchObservedRunningTime="2025-10-04 04:02:45.894201372 +0000 UTC m=+1340.068824575" Oct 04 04:02:45 crc kubenswrapper[4726]: I1004 04:02:45.915870 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.915849208 podStartE2EDuration="2.915849208s" podCreationTimestamp="2025-10-04 04:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:02:45.911492368 +0000 UTC m=+1340.086115581" watchObservedRunningTime="2025-10-04 04:02:45.915849208 +0000 UTC m=+1340.090472421" Oct 04 04:02:46 crc kubenswrapper[4726]: I1004 04:02:46.199741 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:46 crc kubenswrapper[4726]: I1004 04:02:46.199806 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:02:46 crc kubenswrapper[4726]: I1004 04:02:46.898160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerStarted","Data":"d5aa397c8c7499bf062657034bb4a6e79e3a65f258daf3a9748d5e1c5f0afe9d"} Oct 04 04:02:47 crc kubenswrapper[4726]: I1004 04:02:47.907446 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerStarted","Data":"7074e87f33d832b0685f9a537bfec335534441db6dbec0ae99d66e3feb93653c"} Oct 04 04:02:47 crc kubenswrapper[4726]: I1004 04:02:47.908154 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:02:47 crc kubenswrapper[4726]: I1004 04:02:47.934666 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7406374 podStartE2EDuration="6.934358269s" podCreationTimestamp="2025-10-04 04:02:41 +0000 UTC" firstStartedPulling="2025-10-04 04:02:42.804411107 +0000 UTC m=+1336.979034350" lastFinishedPulling="2025-10-04 04:02:46.998131996 +0000 UTC m=+1341.172755219" observedRunningTime="2025-10-04 04:02:47.925130055 +0000 UTC m=+1342.099753268" watchObservedRunningTime="2025-10-04 04:02:47.934358269 +0000 UTC m=+1342.108981482" Oct 04 04:02:48 crc kubenswrapper[4726]: I1004 04:02:48.344940 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:02:49 crc kubenswrapper[4726]: I1004 04:02:49.278998 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:02:51 crc kubenswrapper[4726]: I1004 04:02:51.202166 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:02:51 crc kubenswrapper[4726]: I1004 04:02:51.202649 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:02:51 crc kubenswrapper[4726]: I1004 04:02:51.247425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 04:02:52 crc kubenswrapper[4726]: I1004 04:02:52.255254 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:02:52 crc kubenswrapper[4726]: I1004 04:02:52.255275 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:02:54 crc kubenswrapper[4726]: I1004 04:02:54.272092 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:02:54 crc kubenswrapper[4726]: I1004 04:02:54.272681 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:02:54 crc kubenswrapper[4726]: I1004 04:02:54.278977 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:02:54 crc kubenswrapper[4726]: I1004 04:02:54.321595 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:02:55 crc kubenswrapper[4726]: I1004 04:02:55.002277 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:02:55 crc kubenswrapper[4726]: I1004 04:02:55.355358 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:02:55 crc kubenswrapper[4726]: I1004 04:02:55.355453 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:01 crc kubenswrapper[4726]: I1004 04:03:01.205698 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:03:01 crc kubenswrapper[4726]: I1004 04:03:01.207576 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:03:01 crc kubenswrapper[4726]: I1004 04:03:01.213931 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.061099 4726 generic.go:334] "Generic (PLEG): container finished" podID="ed51a4d9-f248-470f-a887-1ae137e39a66" containerID="d61e64a1a93a2ac275345ec3b2dc40dbce5172981bf6f137d340956567b48a3e" exitCode=137 Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.061163 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed51a4d9-f248-470f-a887-1ae137e39a66","Type":"ContainerDied","Data":"d61e64a1a93a2ac275345ec3b2dc40dbce5172981bf6f137d340956567b48a3e"} Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.061832 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed51a4d9-f248-470f-a887-1ae137e39a66","Type":"ContainerDied","Data":"057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a"} Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.061870 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="057d8bc330afa2e2e4e3a2e1574675b2cb8f7b3652ee1d524a054279da07bb7a" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.071539 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.135916 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.223874 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krp7t\" (UniqueName: \"kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t\") pod \"ed51a4d9-f248-470f-a887-1ae137e39a66\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.224019 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data\") pod \"ed51a4d9-f248-470f-a887-1ae137e39a66\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.224067 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle\") pod \"ed51a4d9-f248-470f-a887-1ae137e39a66\" (UID: \"ed51a4d9-f248-470f-a887-1ae137e39a66\") " Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.234319 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t" (OuterVolumeSpecName: "kube-api-access-krp7t") pod "ed51a4d9-f248-470f-a887-1ae137e39a66" (UID: "ed51a4d9-f248-470f-a887-1ae137e39a66"). InnerVolumeSpecName "kube-api-access-krp7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.266117 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed51a4d9-f248-470f-a887-1ae137e39a66" (UID: "ed51a4d9-f248-470f-a887-1ae137e39a66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.269991 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data" (OuterVolumeSpecName: "config-data") pod "ed51a4d9-f248-470f-a887-1ae137e39a66" (UID: "ed51a4d9-f248-470f-a887-1ae137e39a66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.326266 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.326306 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed51a4d9-f248-470f-a887-1ae137e39a66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:02 crc kubenswrapper[4726]: I1004 04:03:02.326322 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krp7t\" (UniqueName: \"kubernetes.io/projected/ed51a4d9-f248-470f-a887-1ae137e39a66-kube-api-access-krp7t\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.075597 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.117538 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.171954 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.197816 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:03:03 crc kubenswrapper[4726]: E1004 04:03:03.198516 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed51a4d9-f248-470f-a887-1ae137e39a66" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.198638 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed51a4d9-f248-470f-a887-1ae137e39a66" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.199034 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed51a4d9-f248-470f-a887-1ae137e39a66" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.200634 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.203172 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.203416 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.203478 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.218466 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.357711 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.357845 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.358081 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.358215 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.358344 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6gn9\" (UniqueName: \"kubernetes.io/projected/df4fee1d-47f1-48e0-98bd-f4db94f5a543-kube-api-access-m6gn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.460553 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.460700 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6gn9\" (UniqueName: \"kubernetes.io/projected/df4fee1d-47f1-48e0-98bd-f4db94f5a543-kube-api-access-m6gn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.460975 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.461067 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.461227 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.466645 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.466761 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.468014 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.468417 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4fee1d-47f1-48e0-98bd-f4db94f5a543-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.484793 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6gn9\" (UniqueName: \"kubernetes.io/projected/df4fee1d-47f1-48e0-98bd-f4db94f5a543-kube-api-access-m6gn9\") pod \"nova-cell1-novncproxy-0\" (UID: \"df4fee1d-47f1-48e0-98bd-f4db94f5a543\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:03 crc kubenswrapper[4726]: I1004 04:03:03.525929 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.000803 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.090098 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"df4fee1d-47f1-48e0-98bd-f4db94f5a543","Type":"ContainerStarted","Data":"8e205d6543b4cc627e412a4f4eb8e7f0496c0780316e03e92ad48aa85018957c"} Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.187909 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.187980 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.275336 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.276038 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.279619 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.282439 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:03:04 crc kubenswrapper[4726]: I1004 04:03:04.512306 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed51a4d9-f248-470f-a887-1ae137e39a66" path="/var/lib/kubelet/pods/ed51a4d9-f248-470f-a887-1ae137e39a66/volumes" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.107495 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"df4fee1d-47f1-48e0-98bd-f4db94f5a543","Type":"ContainerStarted","Data":"65f6a3e6fabfd888705a775ff9f3cf413b1642b758d7720034d27b4e279a4163"} Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.108160 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.113577 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.190970 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.190937482 podStartE2EDuration="2.190937482s" podCreationTimestamp="2025-10-04 04:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:05.157812218 +0000 UTC m=+1359.332435471" watchObservedRunningTime="2025-10-04 04:03:05.190937482 +0000 UTC m=+1359.365560725" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.402510 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.405005 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.422088 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504320 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504379 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504431 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504462 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504522 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsnk6\" (UniqueName: \"kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.504575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.606534 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.606599 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsnk6\" (UniqueName: \"kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.607580 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.607723 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.607750 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.607770 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.608044 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.608923 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.611404 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.611643 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.611929 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.632341 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsnk6\" (UniqueName: \"kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6\") pod \"dnsmasq-dns-5c7b6c5df9-spvr7\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:05 crc kubenswrapper[4726]: I1004 04:03:05.735266 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:06 crc kubenswrapper[4726]: I1004 04:03:06.199765 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.124397 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerID="de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e" exitCode=0 Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.124461 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" event={"ID":"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9","Type":"ContainerDied","Data":"de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e"} Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.124967 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" event={"ID":"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9","Type":"ContainerStarted","Data":"f44cffa2404e3a34fcfa8ffc472ad52a97f600bba8fe3fc229368bd7af634fa1"} Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.297222 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.297986 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-central-agent" containerID="cri-o://ca6c6ce3b9f0ecfe725e6a57a4de9eb79e470a0e269311612c1cc4afeabb98dd" gracePeriod=30 Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.298024 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="sg-core" containerID="cri-o://d5aa397c8c7499bf062657034bb4a6e79e3a65f258daf3a9748d5e1c5f0afe9d" gracePeriod=30 Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.298041 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-notification-agent" containerID="cri-o://88ce55993e6aaf7ebed96ae8b671f5efa5f400e6ddc564ee5db20bb74ea7f124" gracePeriod=30 Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.298116 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="proxy-httpd" containerID="cri-o://7074e87f33d832b0685f9a537bfec335534441db6dbec0ae99d66e3feb93653c" gracePeriod=30 Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.309496 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": EOF" Oct 04 04:03:07 crc kubenswrapper[4726]: I1004 04:03:07.535418 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.137688 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" event={"ID":"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9","Type":"ContainerStarted","Data":"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821"} Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.137798 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141369 4726 generic.go:334] "Generic (PLEG): container finished" podID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerID="7074e87f33d832b0685f9a537bfec335534441db6dbec0ae99d66e3feb93653c" exitCode=0 Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141397 4726 generic.go:334] "Generic (PLEG): container finished" podID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerID="d5aa397c8c7499bf062657034bb4a6e79e3a65f258daf3a9748d5e1c5f0afe9d" exitCode=2 Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141406 4726 generic.go:334] "Generic (PLEG): container finished" podID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerID="ca6c6ce3b9f0ecfe725e6a57a4de9eb79e470a0e269311612c1cc4afeabb98dd" exitCode=0 Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141421 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerDied","Data":"7074e87f33d832b0685f9a537bfec335534441db6dbec0ae99d66e3feb93653c"} Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141460 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerDied","Data":"d5aa397c8c7499bf062657034bb4a6e79e3a65f258daf3a9748d5e1c5f0afe9d"} Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141473 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerDied","Data":"ca6c6ce3b9f0ecfe725e6a57a4de9eb79e470a0e269311612c1cc4afeabb98dd"} Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141555 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-log" containerID="cri-o://0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca" gracePeriod=30 Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.141599 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-api" containerID="cri-o://8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4" gracePeriod=30 Oct 04 04:03:08 crc kubenswrapper[4726]: I1004 04:03:08.526557 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:09 crc kubenswrapper[4726]: I1004 04:03:09.155731 4726 generic.go:334] "Generic (PLEG): container finished" podID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerID="0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca" exitCode=143 Oct 04 04:03:09 crc kubenswrapper[4726]: I1004 04:03:09.155914 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerDied","Data":"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca"} Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.187886 4726 generic.go:334] "Generic (PLEG): container finished" podID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerID="88ce55993e6aaf7ebed96ae8b671f5efa5f400e6ddc564ee5db20bb74ea7f124" exitCode=0 Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.187947 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerDied","Data":"88ce55993e6aaf7ebed96ae8b671f5efa5f400e6ddc564ee5db20bb74ea7f124"} Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.188701 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0ed3b07c-8a76-4884-96e0-94f9ca3860f1","Type":"ContainerDied","Data":"f253fc1f2e654ca1867a5d40e2751f7f2418732264d2b0d727a511e87d4a23a6"} Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.188730 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f253fc1f2e654ca1867a5d40e2751f7f2418732264d2b0d727a511e87d4a23a6" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.214407 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.241534 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" podStartSLOduration=6.241517692 podStartE2EDuration="6.241517692s" podCreationTimestamp="2025-10-04 04:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:08.164759223 +0000 UTC m=+1362.339382436" watchObservedRunningTime="2025-10-04 04:03:11.241517692 +0000 UTC m=+1365.416140905" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.325937 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326000 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326060 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326087 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326139 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326240 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86nmn\" (UniqueName: \"kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326271 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.326295 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts\") pod \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\" (UID: \"0ed3b07c-8a76-4884-96e0-94f9ca3860f1\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.327657 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.327649 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.338019 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn" (OuterVolumeSpecName: "kube-api-access-86nmn") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "kube-api-access-86nmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.339237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts" (OuterVolumeSpecName: "scripts") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.363068 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.387869 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.423589 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428165 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428194 4726 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428207 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86nmn\" (UniqueName: \"kubernetes.io/projected/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-kube-api-access-86nmn\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428222 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428234 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428245 4726 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.428256 4726 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.461092 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data" (OuterVolumeSpecName: "config-data") pod "0ed3b07c-8a76-4884-96e0-94f9ca3860f1" (UID: "0ed3b07c-8a76-4884-96e0-94f9ca3860f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.530136 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ed3b07c-8a76-4884-96e0-94f9ca3860f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.715048 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.833865 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs\") pod \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.834469 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hddbb\" (UniqueName: \"kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb\") pod \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.834546 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle\") pod \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.834578 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data\") pod \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\" (UID: \"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f\") " Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.835640 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs" (OuterVolumeSpecName: "logs") pod "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" (UID: "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.841515 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb" (OuterVolumeSpecName: "kube-api-access-hddbb") pod "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" (UID: "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f"). InnerVolumeSpecName "kube-api-access-hddbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.861990 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" (UID: "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.869682 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data" (OuterVolumeSpecName: "config-data") pod "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" (UID: "a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.936997 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.937031 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hddbb\" (UniqueName: \"kubernetes.io/projected/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-kube-api-access-hddbb\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.937043 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:11 crc kubenswrapper[4726]: I1004 04:03:11.937052 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200082 4726 generic.go:334] "Generic (PLEG): container finished" podID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerID="8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4" exitCode=0 Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200204 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200663 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200724 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerDied","Data":"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4"} Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200798 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f","Type":"ContainerDied","Data":"70258491387eb9fc9854dd47b76f47daab326706d6a47d5a0a8d9c283ed0e50a"} Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.200819 4726 scope.go:117] "RemoveContainer" containerID="8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.234983 4726 scope.go:117] "RemoveContainer" containerID="0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.269418 4726 scope.go:117] "RemoveContainer" containerID="8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.269557 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.271235 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4\": container with ID starting with 8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4 not found: ID does not exist" containerID="8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.271283 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4"} err="failed to get container status \"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4\": rpc error: code = NotFound desc = could not find container \"8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4\": container with ID starting with 8785fc0b68c08f4417ee9b81ecb45010b7a3cb99253acfd93d19d893d1c151e4 not found: ID does not exist" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.271309 4726 scope.go:117] "RemoveContainer" containerID="0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.272859 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca\": container with ID starting with 0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca not found: ID does not exist" containerID="0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.272880 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca"} err="failed to get container status \"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca\": rpc error: code = NotFound desc = could not find container \"0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca\": container with ID starting with 0112cfc4a60b5ce285b1a92713073a71f7a843efe9d67fc835876ea535771aca not found: ID does not exist" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.293156 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.308175 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.317482 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327156 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327549 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="proxy-httpd" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327564 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="proxy-httpd" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327585 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-notification-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327592 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-notification-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327603 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-central-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327609 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-central-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327624 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-log" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327630 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-log" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327644 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-api" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327650 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-api" Oct 04 04:03:12 crc kubenswrapper[4726]: E1004 04:03:12.327659 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="sg-core" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327664 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="sg-core" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327821 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-api" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327837 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-central-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327844 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="ceilometer-notification-agent" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327854 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="sg-core" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327860 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" containerName="nova-api-log" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.327879 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" containerName="proxy-httpd" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.329495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.335482 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.335756 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.335896 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.343159 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.345395 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.350930 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.351165 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.351283 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.374125 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.389586 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447097 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-run-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447250 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c4gc\" (UniqueName: \"kubernetes.io/projected/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-kube-api-access-6c4gc\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447278 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-config-data\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447300 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-log-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447335 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsjll\" (UniqueName: \"kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447387 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447405 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-scripts\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447899 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447963 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.447984 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.448031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.512971 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed3b07c-8a76-4884-96e0-94f9ca3860f1" path="/var/lib/kubelet/pods/0ed3b07c-8a76-4884-96e0-94f9ca3860f1/volumes" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.513705 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f" path="/var/lib/kubelet/pods/a4fb1f7b-e26c-4f9d-b3aa-efa89e23c03f/volumes" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549531 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-scripts\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549575 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549630 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549669 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549694 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-run-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549788 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549817 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c4gc\" (UniqueName: \"kubernetes.io/projected/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-kube-api-access-6c4gc\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549840 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-config-data\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549855 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-log-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549881 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsjll\" (UniqueName: \"kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549901 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.549918 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.550941 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-run-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.551060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.551233 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-log-httpd\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.553911 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-scripts\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.553995 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-config-data\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.554418 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.554951 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.556385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.556807 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.557450 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.558362 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.560764 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.569382 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsjll\" (UniqueName: \"kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll\") pod \"nova-api-0\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.570183 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c4gc\" (UniqueName: \"kubernetes.io/projected/00a3dd9a-1341-4141-a0e6-8dc5b3e73565-kube-api-access-6c4gc\") pod \"ceilometer-0\" (UID: \"00a3dd9a-1341-4141-a0e6-8dc5b3e73565\") " pod="openstack/ceilometer-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.657262 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:12 crc kubenswrapper[4726]: I1004 04:03:12.671247 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:03:13 crc kubenswrapper[4726]: I1004 04:03:13.204869 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:03:13 crc kubenswrapper[4726]: W1004 04:03:13.209193 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a3dd9a_1341_4141_a0e6_8dc5b3e73565.slice/crio-e5a181bdf592bbd9f8db32c3976f44455003822c02ed2376d0f48299dae82bfc WatchSource:0}: Error finding container e5a181bdf592bbd9f8db32c3976f44455003822c02ed2376d0f48299dae82bfc: Status 404 returned error can't find the container with id e5a181bdf592bbd9f8db32c3976f44455003822c02ed2376d0f48299dae82bfc Oct 04 04:03:13 crc kubenswrapper[4726]: I1004 04:03:13.214247 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:03:13 crc kubenswrapper[4726]: I1004 04:03:13.244253 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:13 crc kubenswrapper[4726]: W1004 04:03:13.249900 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4604160e_5238_4a36_bb4e_2a21b2cbc798.slice/crio-7cf10348a5071b15c12dab12312b3ad990b840c17170d0c6db4b39b08e2f1b78 WatchSource:0}: Error finding container 7cf10348a5071b15c12dab12312b3ad990b840c17170d0c6db4b39b08e2f1b78: Status 404 returned error can't find the container with id 7cf10348a5071b15c12dab12312b3ad990b840c17170d0c6db4b39b08e2f1b78 Oct 04 04:03:13 crc kubenswrapper[4726]: I1004 04:03:13.526540 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:13 crc kubenswrapper[4726]: I1004 04:03:13.559191 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.230927 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00a3dd9a-1341-4141-a0e6-8dc5b3e73565","Type":"ContainerStarted","Data":"1268f2438747f6f48ff49113c7f25fe1171323c44abd67b4c500354bbc610f44"} Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.230994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00a3dd9a-1341-4141-a0e6-8dc5b3e73565","Type":"ContainerStarted","Data":"e5a181bdf592bbd9f8db32c3976f44455003822c02ed2376d0f48299dae82bfc"} Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.234222 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerStarted","Data":"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6"} Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.234250 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerStarted","Data":"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e"} Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.234282 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerStarted","Data":"7cf10348a5071b15c12dab12312b3ad990b840c17170d0c6db4b39b08e2f1b78"} Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.251882 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.260749 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.260723156 podStartE2EDuration="2.260723156s" podCreationTimestamp="2025-10-04 04:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:14.251449698 +0000 UTC m=+1368.426072911" watchObservedRunningTime="2025-10-04 04:03:14.260723156 +0000 UTC m=+1368.435346369" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.405265 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gbbrr"] Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.406731 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.408702 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.408866 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.413906 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gbbrr"] Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.485697 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.485828 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.485846 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.486037 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5kvf\" (UniqueName: \"kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.588326 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.588412 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.588430 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.588481 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5kvf\" (UniqueName: \"kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.592388 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.592586 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.599614 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.602890 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5kvf\" (UniqueName: \"kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf\") pod \"nova-cell1-cell-mapping-gbbrr\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:14 crc kubenswrapper[4726]: I1004 04:03:14.723377 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.183365 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gbbrr"] Oct 04 04:03:15 crc kubenswrapper[4726]: W1004 04:03:15.192849 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd02f9ea_7639_46f4_b330_8a5cb3864e9d.slice/crio-8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb WatchSource:0}: Error finding container 8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb: Status 404 returned error can't find the container with id 8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.249071 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gbbrr" event={"ID":"cd02f9ea-7639-46f4-b330-8a5cb3864e9d","Type":"ContainerStarted","Data":"8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb"} Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.251841 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00a3dd9a-1341-4141-a0e6-8dc5b3e73565","Type":"ContainerStarted","Data":"6ef69265e26d82acf3c5f4307202cc7c4c3e53d48e0376be4b63b37e9bcf0926"} Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.737551 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.811926 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:03:15 crc kubenswrapper[4726]: I1004 04:03:15.812207 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="dnsmasq-dns" containerID="cri-o://de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5" gracePeriod=10 Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.253538 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.266181 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00a3dd9a-1341-4141-a0e6-8dc5b3e73565","Type":"ContainerStarted","Data":"a3b3b832522804602229cff0bbb79e27c08dcf77a25fdd6ecc061817a9cc267e"} Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.269209 4726 generic.go:334] "Generic (PLEG): container finished" podID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerID="de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5" exitCode=0 Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.269263 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" event={"ID":"7f4503d8-cfb2-47db-8c7d-597566f57dfa","Type":"ContainerDied","Data":"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5"} Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.269284 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" event={"ID":"7f4503d8-cfb2-47db-8c7d-597566f57dfa","Type":"ContainerDied","Data":"9492396cc852e7ede59792faae30074dff0a7fdfdb1e531a313918a73d53a596"} Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.269300 4726 scope.go:117] "RemoveContainer" containerID="de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.269318 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-9nfgv" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.276139 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gbbrr" event={"ID":"cd02f9ea-7639-46f4-b330-8a5cb3864e9d","Type":"ContainerStarted","Data":"624afb6db3583fb9ac26f7e2920aa28fced0811676a72d315ff1c88c5ac52b21"} Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.293981 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gbbrr" podStartSLOduration=2.293964803 podStartE2EDuration="2.293964803s" podCreationTimestamp="2025-10-04 04:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:16.290650384 +0000 UTC m=+1370.465273617" watchObservedRunningTime="2025-10-04 04:03:16.293964803 +0000 UTC m=+1370.468588016" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.304969 4726 scope.go:117] "RemoveContainer" containerID="3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.333959 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z674z\" (UniqueName: \"kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.334052 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.334171 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.334226 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.334251 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.334281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb\") pod \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\" (UID: \"7f4503d8-cfb2-47db-8c7d-597566f57dfa\") " Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.339365 4726 scope.go:117] "RemoveContainer" containerID="de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.339761 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z" (OuterVolumeSpecName: "kube-api-access-z674z") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "kube-api-access-z674z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: E1004 04:03:16.350741 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5\": container with ID starting with de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5 not found: ID does not exist" containerID="de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.350786 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5"} err="failed to get container status \"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5\": rpc error: code = NotFound desc = could not find container \"de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5\": container with ID starting with de8cbf8bcf70201a95d156a2d721029181f4e93b721108a0ce14dc091e701df5 not found: ID does not exist" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.350819 4726 scope.go:117] "RemoveContainer" containerID="3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9" Oct 04 04:03:16 crc kubenswrapper[4726]: E1004 04:03:16.351135 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9\": container with ID starting with 3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9 not found: ID does not exist" containerID="3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.351166 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9"} err="failed to get container status \"3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9\": rpc error: code = NotFound desc = could not find container \"3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9\": container with ID starting with 3440cf6fe51a3eaa449ed9867d2cb8a0dda4df5f35ac60b77fc3b2892aa81de9 not found: ID does not exist" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.389822 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config" (OuterVolumeSpecName: "config") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.393375 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.397187 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.397599 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.400835 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f4503d8-cfb2-47db-8c7d-597566f57dfa" (UID: "7f4503d8-cfb2-47db-8c7d-597566f57dfa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436056 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z674z\" (UniqueName: \"kubernetes.io/projected/7f4503d8-cfb2-47db-8c7d-597566f57dfa-kube-api-access-z674z\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436086 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436096 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436118 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436127 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.436135 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4503d8-cfb2-47db-8c7d-597566f57dfa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.599526 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:03:16 crc kubenswrapper[4726]: I1004 04:03:16.607244 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-9nfgv"] Oct 04 04:03:17 crc kubenswrapper[4726]: I1004 04:03:17.287233 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"00a3dd9a-1341-4141-a0e6-8dc5b3e73565","Type":"ContainerStarted","Data":"a692f4d46c6ec72eeabacfa54c1d60548993b6827ab4b721f5a572305b355d9a"} Oct 04 04:03:17 crc kubenswrapper[4726]: I1004 04:03:17.287736 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:03:17 crc kubenswrapper[4726]: I1004 04:03:17.317940 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.619444659 podStartE2EDuration="5.317920552s" podCreationTimestamp="2025-10-04 04:03:12 +0000 UTC" firstStartedPulling="2025-10-04 04:03:13.214026319 +0000 UTC m=+1367.388649532" lastFinishedPulling="2025-10-04 04:03:16.912502212 +0000 UTC m=+1371.087125425" observedRunningTime="2025-10-04 04:03:17.303625021 +0000 UTC m=+1371.478248234" watchObservedRunningTime="2025-10-04 04:03:17.317920552 +0000 UTC m=+1371.492543755" Oct 04 04:03:18 crc kubenswrapper[4726]: I1004 04:03:18.515294 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" path="/var/lib/kubelet/pods/7f4503d8-cfb2-47db-8c7d-597566f57dfa/volumes" Oct 04 04:03:20 crc kubenswrapper[4726]: I1004 04:03:20.322048 4726 generic.go:334] "Generic (PLEG): container finished" podID="cd02f9ea-7639-46f4-b330-8a5cb3864e9d" containerID="624afb6db3583fb9ac26f7e2920aa28fced0811676a72d315ff1c88c5ac52b21" exitCode=0 Oct 04 04:03:20 crc kubenswrapper[4726]: I1004 04:03:20.322170 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gbbrr" event={"ID":"cd02f9ea-7639-46f4-b330-8a5cb3864e9d","Type":"ContainerDied","Data":"624afb6db3583fb9ac26f7e2920aa28fced0811676a72d315ff1c88c5ac52b21"} Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.727948 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.845057 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts\") pod \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.845198 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data\") pod \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.845410 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5kvf\" (UniqueName: \"kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf\") pod \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.845458 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle\") pod \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\" (UID: \"cd02f9ea-7639-46f4-b330-8a5cb3864e9d\") " Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.851759 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts" (OuterVolumeSpecName: "scripts") pod "cd02f9ea-7639-46f4-b330-8a5cb3864e9d" (UID: "cd02f9ea-7639-46f4-b330-8a5cb3864e9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.857334 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf" (OuterVolumeSpecName: "kube-api-access-f5kvf") pod "cd02f9ea-7639-46f4-b330-8a5cb3864e9d" (UID: "cd02f9ea-7639-46f4-b330-8a5cb3864e9d"). InnerVolumeSpecName "kube-api-access-f5kvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.881952 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd02f9ea-7639-46f4-b330-8a5cb3864e9d" (UID: "cd02f9ea-7639-46f4-b330-8a5cb3864e9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.892929 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data" (OuterVolumeSpecName: "config-data") pod "cd02f9ea-7639-46f4-b330-8a5cb3864e9d" (UID: "cd02f9ea-7639-46f4-b330-8a5cb3864e9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.950293 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5kvf\" (UniqueName: \"kubernetes.io/projected/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-kube-api-access-f5kvf\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.950328 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.950337 4726 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:21 crc kubenswrapper[4726]: I1004 04:03:21.950346 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd02f9ea-7639-46f4-b330-8a5cb3864e9d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.346383 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gbbrr" event={"ID":"cd02f9ea-7639-46f4-b330-8a5cb3864e9d","Type":"ContainerDied","Data":"8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb"} Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.346445 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bed5aeb23ddf29e328b8553c08b4f58e113053fc5e0a56ceaab3e116c9127cb" Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.346460 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gbbrr" Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.557524 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.557784 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerName="nova-scheduler-scheduler" containerID="cri-o://1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" gracePeriod=30 Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.568379 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.568659 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-log" containerID="cri-o://1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" gracePeriod=30 Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.568810 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-api" containerID="cri-o://e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" gracePeriod=30 Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.627777 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.628238 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" containerID="cri-o://c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a" gracePeriod=30 Oct 04 04:03:22 crc kubenswrapper[4726]: I1004 04:03:22.628413 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" containerID="cri-o://5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156" gracePeriod=30 Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.166619 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.273958 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274034 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274171 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274201 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274257 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsjll\" (UniqueName: \"kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll\") pod \"4604160e-5238-4a36-bb4e-2a21b2cbc798\" (UID: \"4604160e-5238-4a36-bb4e-2a21b2cbc798\") " Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.274709 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs" (OuterVolumeSpecName: "logs") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.280566 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll" (OuterVolumeSpecName: "kube-api-access-dsjll") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "kube-api-access-dsjll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.307643 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.309017 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data" (OuterVolumeSpecName: "config-data") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.344070 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.362191 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4604160e-5238-4a36-bb4e-2a21b2cbc798" (UID: "4604160e-5238-4a36-bb4e-2a21b2cbc798"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.372636 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerID="c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a" exitCode=143 Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.372701 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerDied","Data":"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a"} Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.374726 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.374763 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerDied","Data":"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6"} Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.374820 4726 scope.go:117] "RemoveContainer" containerID="e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.374676 4726 generic.go:334] "Generic (PLEG): container finished" podID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerID="e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" exitCode=0 Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.375885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerDied","Data":"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e"} Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376174 4726 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376201 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376244 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4604160e-5238-4a36-bb4e-2a21b2cbc798-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376254 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsjll\" (UniqueName: \"kubernetes.io/projected/4604160e-5238-4a36-bb4e-2a21b2cbc798-kube-api-access-dsjll\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376266 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376274 4726 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4604160e-5238-4a36-bb4e-2a21b2cbc798-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.376943 4726 generic.go:334] "Generic (PLEG): container finished" podID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerID="1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" exitCode=143 Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.377340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4604160e-5238-4a36-bb4e-2a21b2cbc798","Type":"ContainerDied","Data":"7cf10348a5071b15c12dab12312b3ad990b840c17170d0c6db4b39b08e2f1b78"} Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.402495 4726 scope.go:117] "RemoveContainer" containerID="1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.432269 4726 scope.go:117] "RemoveContainer" containerID="e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.444194 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.444774 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6\": container with ID starting with e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6 not found: ID does not exist" containerID="e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.445256 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6"} err="failed to get container status \"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6\": rpc error: code = NotFound desc = could not find container \"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6\": container with ID starting with e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6 not found: ID does not exist" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.445291 4726 scope.go:117] "RemoveContainer" containerID="1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.450258 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e\": container with ID starting with 1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e not found: ID does not exist" containerID="1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.450326 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e"} err="failed to get container status \"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e\": rpc error: code = NotFound desc = could not find container \"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e\": container with ID starting with 1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e not found: ID does not exist" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.450383 4726 scope.go:117] "RemoveContainer" containerID="e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.451945 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6"} err="failed to get container status \"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6\": rpc error: code = NotFound desc = could not find container \"e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6\": container with ID starting with e7c0463b46adccf568df2a0d376064d7c657d14a267796ad0b9eac125a9fa5d6 not found: ID does not exist" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.451997 4726 scope.go:117] "RemoveContainer" containerID="1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.452402 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e"} err="failed to get container status \"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e\": rpc error: code = NotFound desc = could not find container \"1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e\": container with ID starting with 1f854e93abeabb98ca92856931e0d1cf6ca061bca248cb49e5a4539503468c1e not found: ID does not exist" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.458258 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467124 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.467700 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-log" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467719 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-log" Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.467741 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-api" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467750 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-api" Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.467774 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="init" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467783 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="init" Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.467800 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd02f9ea-7639-46f4-b330-8a5cb3864e9d" containerName="nova-manage" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467808 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd02f9ea-7639-46f4-b330-8a5cb3864e9d" containerName="nova-manage" Oct 04 04:03:23 crc kubenswrapper[4726]: E1004 04:03:23.467827 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="dnsmasq-dns" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.467834 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="dnsmasq-dns" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.468094 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd02f9ea-7639-46f4-b330-8a5cb3864e9d" containerName="nova-manage" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.468129 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4503d8-cfb2-47db-8c7d-597566f57dfa" containerName="dnsmasq-dns" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.468152 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-api" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.468166 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" containerName="nova-api-log" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.469428 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.473661 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.473835 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.474366 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.475071 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583593 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flntn\" (UniqueName: \"kubernetes.io/projected/5c95d78d-1f21-48f0-903e-713cde259066-kube-api-access-flntn\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583657 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-config-data\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583701 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c95d78d-1f21-48f0-903e-713cde259066-logs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583752 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-public-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583838 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.583884 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.685826 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c95d78d-1f21-48f0-903e-713cde259066-logs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.685910 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-public-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.686012 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.686053 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.686123 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flntn\" (UniqueName: \"kubernetes.io/projected/5c95d78d-1f21-48f0-903e-713cde259066-kube-api-access-flntn\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.686165 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-config-data\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.686604 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c95d78d-1f21-48f0-903e-713cde259066-logs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.690260 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-public-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.690476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-config-data\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.691791 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.692571 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c95d78d-1f21-48f0-903e-713cde259066-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.710734 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flntn\" (UniqueName: \"kubernetes.io/projected/5c95d78d-1f21-48f0-903e-713cde259066-kube-api-access-flntn\") pod \"nova-api-0\" (UID: \"5c95d78d-1f21-48f0-903e-713cde259066\") " pod="openstack/nova-api-0" Oct 04 04:03:23 crc kubenswrapper[4726]: I1004 04:03:23.787466 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:03:24 crc kubenswrapper[4726]: E1004 04:03:24.282838 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:03:24 crc kubenswrapper[4726]: E1004 04:03:24.285379 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:03:24 crc kubenswrapper[4726]: I1004 04:03:24.287159 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:03:24 crc kubenswrapper[4726]: E1004 04:03:24.287367 4726 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:03:24 crc kubenswrapper[4726]: E1004 04:03:24.287564 4726 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerName="nova-scheduler-scheduler" Oct 04 04:03:24 crc kubenswrapper[4726]: I1004 04:03:24.388642 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c95d78d-1f21-48f0-903e-713cde259066","Type":"ContainerStarted","Data":"7826f33905f67c70ebe681354513b8c0e275c975eafd80afe1dac3ce3bdbb46c"} Oct 04 04:03:24 crc kubenswrapper[4726]: I1004 04:03:24.513157 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4604160e-5238-4a36-bb4e-2a21b2cbc798" path="/var/lib/kubelet/pods/4604160e-5238-4a36-bb4e-2a21b2cbc798/volumes" Oct 04 04:03:25 crc kubenswrapper[4726]: I1004 04:03:25.412846 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c95d78d-1f21-48f0-903e-713cde259066","Type":"ContainerStarted","Data":"c44514111f8ddc27aa9b4ebdf6b2922adfb3058dbb966c15ff63f82ecbacd3bb"} Oct 04 04:03:25 crc kubenswrapper[4726]: I1004 04:03:25.412918 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5c95d78d-1f21-48f0-903e-713cde259066","Type":"ContainerStarted","Data":"06cca28b61c3ae37acf100be19f9650fd318d737b386bfa4a2600f4b95a8cfa6"} Oct 04 04:03:25 crc kubenswrapper[4726]: I1004 04:03:25.448823 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.448792635 podStartE2EDuration="2.448792635s" podCreationTimestamp="2025-10-04 04:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:25.435814249 +0000 UTC m=+1379.610437482" watchObservedRunningTime="2025-10-04 04:03:25.448792635 +0000 UTC m=+1379.623415888" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.333138 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.430692 4726 generic.go:334] "Generic (PLEG): container finished" podID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerID="5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156" exitCode=0 Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.430769 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerDied","Data":"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156"} Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.430814 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.430854 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a5ff588-b7a1-40c3-ade4-853a0be28379","Type":"ContainerDied","Data":"12330fae995ea4f279ad2b25c805fd81984dc49ea80394ea6d917bc471fdfced"} Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.430881 4726 scope.go:117] "RemoveContainer" containerID="5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.452429 4726 scope.go:117] "RemoveContainer" containerID="c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461416 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs\") pod \"6a5ff588-b7a1-40c3-ade4-853a0be28379\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461549 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs\") pod \"6a5ff588-b7a1-40c3-ade4-853a0be28379\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461696 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data\") pod \"6a5ff588-b7a1-40c3-ade4-853a0be28379\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461753 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle\") pod \"6a5ff588-b7a1-40c3-ade4-853a0be28379\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461779 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m227\" (UniqueName: \"kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227\") pod \"6a5ff588-b7a1-40c3-ade4-853a0be28379\" (UID: \"6a5ff588-b7a1-40c3-ade4-853a0be28379\") " Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.461907 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs" (OuterVolumeSpecName: "logs") pod "6a5ff588-b7a1-40c3-ade4-853a0be28379" (UID: "6a5ff588-b7a1-40c3-ade4-853a0be28379"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.462307 4726 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5ff588-b7a1-40c3-ade4-853a0be28379-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.491313 4726 scope.go:117] "RemoveContainer" containerID="5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156" Oct 04 04:03:26 crc kubenswrapper[4726]: E1004 04:03:26.491926 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156\": container with ID starting with 5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156 not found: ID does not exist" containerID="5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.491956 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156"} err="failed to get container status \"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156\": rpc error: code = NotFound desc = could not find container \"5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156\": container with ID starting with 5f8031f32153bec8e6894bad94196d34fe7528181f7e3a9db7c8b276afeff156 not found: ID does not exist" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.491980 4726 scope.go:117] "RemoveContainer" containerID="c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a" Oct 04 04:03:26 crc kubenswrapper[4726]: E1004 04:03:26.493027 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a\": container with ID starting with c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a not found: ID does not exist" containerID="c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.493048 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a"} err="failed to get container status \"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a\": rpc error: code = NotFound desc = could not find container \"c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a\": container with ID starting with c0fe7618a628e53abbff55706ea6a1f124dd15de1dbda44bb3aabe8c2ee03a1a not found: ID does not exist" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.502470 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227" (OuterVolumeSpecName: "kube-api-access-4m227") pod "6a5ff588-b7a1-40c3-ade4-853a0be28379" (UID: "6a5ff588-b7a1-40c3-ade4-853a0be28379"). InnerVolumeSpecName "kube-api-access-4m227". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.517906 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data" (OuterVolumeSpecName: "config-data") pod "6a5ff588-b7a1-40c3-ade4-853a0be28379" (UID: "6a5ff588-b7a1-40c3-ade4-853a0be28379"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.524089 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a5ff588-b7a1-40c3-ade4-853a0be28379" (UID: "6a5ff588-b7a1-40c3-ade4-853a0be28379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.557901 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6a5ff588-b7a1-40c3-ade4-853a0be28379" (UID: "6a5ff588-b7a1-40c3-ade4-853a0be28379"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.566801 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.566829 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.566840 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5ff588-b7a1-40c3-ade4-853a0be28379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.567022 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m227\" (UniqueName: \"kubernetes.io/projected/6a5ff588-b7a1-40c3-ade4-853a0be28379-kube-api-access-4m227\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.771848 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.779631 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.796007 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:26 crc kubenswrapper[4726]: E1004 04:03:26.796372 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.796388 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" Oct 04 04:03:26 crc kubenswrapper[4726]: E1004 04:03:26.796433 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.796441 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.796615 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.796640 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.797715 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.805803 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.806020 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.819373 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.876684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spx2d\" (UniqueName: \"kubernetes.io/projected/1f13faed-0be2-4916-80ab-f9d7c7d196a3-kube-api-access-spx2d\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.876732 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.876826 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f13faed-0be2-4916-80ab-f9d7c7d196a3-logs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.876870 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.876887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-config-data\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.978806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spx2d\" (UniqueName: \"kubernetes.io/projected/1f13faed-0be2-4916-80ab-f9d7c7d196a3-kube-api-access-spx2d\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.979079 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.979272 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f13faed-0be2-4916-80ab-f9d7c7d196a3-logs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.979372 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.979460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-config-data\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.979943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f13faed-0be2-4916-80ab-f9d7c7d196a3-logs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.983785 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-config-data\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.984240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:26 crc kubenswrapper[4726]: I1004 04:03:26.984864 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f13faed-0be2-4916-80ab-f9d7c7d196a3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:27 crc kubenswrapper[4726]: I1004 04:03:27.005457 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spx2d\" (UniqueName: \"kubernetes.io/projected/1f13faed-0be2-4916-80ab-f9d7c7d196a3-kube-api-access-spx2d\") pod \"nova-metadata-0\" (UID: \"1f13faed-0be2-4916-80ab-f9d7c7d196a3\") " pod="openstack/nova-metadata-0" Oct 04 04:03:27 crc kubenswrapper[4726]: I1004 04:03:27.120805 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:03:27 crc kubenswrapper[4726]: I1004 04:03:27.640160 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.366795 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.472529 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f13faed-0be2-4916-80ab-f9d7c7d196a3","Type":"ContainerStarted","Data":"bd9ee1ffe924e86d0fae9a69c575da9e9f878475a8ecec2c470f1494a2c8195c"} Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.472592 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f13faed-0be2-4916-80ab-f9d7c7d196a3","Type":"ContainerStarted","Data":"f8cccefb871b6a6d862983391bc7fbe2d6ac62e335382afffdd1ad98ae864174"} Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.472610 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1f13faed-0be2-4916-80ab-f9d7c7d196a3","Type":"ContainerStarted","Data":"aecf4420b8d891172ca5e5cb478d613729493843f9ef7247b3a0f0c1ca0af29f"} Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.475137 4726 generic.go:334] "Generic (PLEG): container finished" podID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" exitCode=0 Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.475160 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.475181 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3","Type":"ContainerDied","Data":"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b"} Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.475219 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3","Type":"ContainerDied","Data":"a341266fdc9cd9ee5731ff987226d13947986a107a22d2c6a2bd9f315d3261a3"} Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.475236 4726 scope.go:117] "RemoveContainer" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.507526 4726 scope.go:117] "RemoveContainer" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" Oct 04 04:03:28 crc kubenswrapper[4726]: E1004 04:03:28.507886 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b\": container with ID starting with 1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b not found: ID does not exist" containerID="1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.507917 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b"} err="failed to get container status \"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b\": rpc error: code = NotFound desc = could not find container \"1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b\": container with ID starting with 1ae361db72872eaf19e4d7f5dbf08c7ee225193273fa0c2165d1d134940e1e4b not found: ID does not exist" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.508865 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.508840419 podStartE2EDuration="2.508840419s" podCreationTimestamp="2025-10-04 04:03:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:28.49466421 +0000 UTC m=+1382.669287423" watchObservedRunningTime="2025-10-04 04:03:28.508840419 +0000 UTC m=+1382.683463652" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.510220 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qwkl\" (UniqueName: \"kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl\") pod \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.510447 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data\") pod \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.511238 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle\") pod \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\" (UID: \"e20838cb-bc7d-4917-b2fa-e2a1825b3ba3\") " Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.517306 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl" (OuterVolumeSpecName: "kube-api-access-9qwkl") pod "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" (UID: "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3"). InnerVolumeSpecName "kube-api-access-9qwkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.523378 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" path="/var/lib/kubelet/pods/6a5ff588-b7a1-40c3-ade4-853a0be28379/volumes" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.540675 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" (UID: "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.544016 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data" (OuterVolumeSpecName: "config-data") pod "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" (UID: "e20838cb-bc7d-4917-b2fa-e2a1825b3ba3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.614171 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qwkl\" (UniqueName: \"kubernetes.io/projected/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-kube-api-access-9qwkl\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.614224 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.614243 4726 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.820309 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.843679 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.851058 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:28 crc kubenswrapper[4726]: E1004 04:03:28.851500 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerName="nova-scheduler-scheduler" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.851522 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerName="nova-scheduler-scheduler" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.851746 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" containerName="nova-scheduler-scheduler" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.852516 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.856431 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:03:28 crc kubenswrapper[4726]: I1004 04:03:28.862704 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.019444 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q56f\" (UniqueName: \"kubernetes.io/projected/2a84708a-3dbc-477f-b151-d68b83b4617b-kube-api-access-6q56f\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.019552 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.019599 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-config-data\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.121576 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.121670 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-config-data\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.121760 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q56f\" (UniqueName: \"kubernetes.io/projected/2a84708a-3dbc-477f-b151-d68b83b4617b-kube-api-access-6q56f\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.126471 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-config-data\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.128936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a84708a-3dbc-477f-b151-d68b83b4617b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.151387 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q56f\" (UniqueName: \"kubernetes.io/projected/2a84708a-3dbc-477f-b151-d68b83b4617b-kube-api-access-6q56f\") pod \"nova-scheduler-0\" (UID: \"2a84708a-3dbc-477f-b151-d68b83b4617b\") " pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.217952 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:03:29 crc kubenswrapper[4726]: W1004 04:03:29.727153 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a84708a_3dbc_477f_b151_d68b83b4617b.slice/crio-1035aa557c29a654b91827249e0d50793f0bdcf5831556acf1c15219488fd88a WatchSource:0}: Error finding container 1035aa557c29a654b91827249e0d50793f0bdcf5831556acf1c15219488fd88a: Status 404 returned error can't find the container with id 1035aa557c29a654b91827249e0d50793f0bdcf5831556acf1c15219488fd88a Oct 04 04:03:29 crc kubenswrapper[4726]: I1004 04:03:29.728260 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:03:30 crc kubenswrapper[4726]: I1004 04:03:30.499216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a84708a-3dbc-477f-b151-d68b83b4617b","Type":"ContainerStarted","Data":"439a2c0b2672a211457b0838c9fc27db5e1c32f00041e4492bf6f13d374a1342"} Oct 04 04:03:30 crc kubenswrapper[4726]: I1004 04:03:30.500099 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a84708a-3dbc-477f-b151-d68b83b4617b","Type":"ContainerStarted","Data":"1035aa557c29a654b91827249e0d50793f0bdcf5831556acf1c15219488fd88a"} Oct 04 04:03:30 crc kubenswrapper[4726]: I1004 04:03:30.527820 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.527787814 podStartE2EDuration="2.527787814s" podCreationTimestamp="2025-10-04 04:03:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:03:30.51489748 +0000 UTC m=+1384.689520733" watchObservedRunningTime="2025-10-04 04:03:30.527787814 +0000 UTC m=+1384.702411057" Oct 04 04:03:30 crc kubenswrapper[4726]: I1004 04:03:30.532464 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e20838cb-bc7d-4917-b2fa-e2a1825b3ba3" path="/var/lib/kubelet/pods/e20838cb-bc7d-4917-b2fa-e2a1825b3ba3/volumes" Oct 04 04:03:31 crc kubenswrapper[4726]: I1004 04:03:31.201139 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: i/o timeout" Oct 04 04:03:31 crc kubenswrapper[4726]: I1004 04:03:31.201150 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6a5ff588-b7a1-40c3-ade4-853a0be28379" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: i/o timeout (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:32 crc kubenswrapper[4726]: I1004 04:03:32.121346 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:03:32 crc kubenswrapper[4726]: I1004 04:03:32.122250 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:03:33 crc kubenswrapper[4726]: I1004 04:03:33.788203 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:03:33 crc kubenswrapper[4726]: I1004 04:03:33.788728 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.188283 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.188362 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.188421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.189523 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.189632 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af" gracePeriod=600 Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.218623 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.554855 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af" exitCode=0 Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.554890 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af"} Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.555583 4726 scope.go:117] "RemoveContainer" containerID="aa47dbe894bafb1405e43b89ae09ceb74647b8bb20ef6560ea1ecddadfe0c5ea" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.807244 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5c95d78d-1f21-48f0-903e-713cde259066" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:34 crc kubenswrapper[4726]: I1004 04:03:34.807337 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5c95d78d-1f21-48f0-903e-713cde259066" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:35 crc kubenswrapper[4726]: I1004 04:03:35.569578 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9"} Oct 04 04:03:37 crc kubenswrapper[4726]: I1004 04:03:37.121559 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:03:37 crc kubenswrapper[4726]: I1004 04:03:37.121955 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:03:38 crc kubenswrapper[4726]: I1004 04:03:38.135335 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f13faed-0be2-4916-80ab-f9d7c7d196a3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:38 crc kubenswrapper[4726]: I1004 04:03:38.135358 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1f13faed-0be2-4916-80ab-f9d7c7d196a3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:03:39 crc kubenswrapper[4726]: I1004 04:03:39.218263 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:03:39 crc kubenswrapper[4726]: I1004 04:03:39.271671 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:03:39 crc kubenswrapper[4726]: I1004 04:03:39.639671 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:03:42 crc kubenswrapper[4726]: I1004 04:03:42.686161 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:03:43 crc kubenswrapper[4726]: I1004 04:03:43.795303 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:03:43 crc kubenswrapper[4726]: I1004 04:03:43.798131 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:03:43 crc kubenswrapper[4726]: I1004 04:03:43.801045 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:03:43 crc kubenswrapper[4726]: I1004 04:03:43.802555 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:03:44 crc kubenswrapper[4726]: I1004 04:03:44.671678 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:03:44 crc kubenswrapper[4726]: I1004 04:03:44.681000 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:03:47 crc kubenswrapper[4726]: I1004 04:03:47.134708 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:03:47 crc kubenswrapper[4726]: I1004 04:03:47.140917 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:03:47 crc kubenswrapper[4726]: I1004 04:03:47.142377 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:03:47 crc kubenswrapper[4726]: I1004 04:03:47.709486 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:03:49 crc kubenswrapper[4726]: I1004 04:03:49.944596 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:03:49 crc kubenswrapper[4726]: I1004 04:03:49.950064 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:49 crc kubenswrapper[4726]: I1004 04:03:49.957024 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.030230 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfm4\" (UniqueName: \"kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.030590 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.030745 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.133405 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.133887 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.133912 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.134169 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.134216 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfm4\" (UniqueName: \"kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.164963 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfm4\" (UniqueName: \"kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4\") pod \"community-operators-vpqb5\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:50 crc kubenswrapper[4726]: I1004 04:03:50.281774 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:03:51 crc kubenswrapper[4726]: I1004 04:03:50.858810 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:03:51 crc kubenswrapper[4726]: W1004 04:03:50.860361 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdf7f8c5_c25a_49c3_9628_81bcc798af5e.slice/crio-324a30ecf3b4be1667196022de6992ccb11b383614d5e197450a49d29864fb32 WatchSource:0}: Error finding container 324a30ecf3b4be1667196022de6992ccb11b383614d5e197450a49d29864fb32: Status 404 returned error can't find the container with id 324a30ecf3b4be1667196022de6992ccb11b383614d5e197450a49d29864fb32 Oct 04 04:03:51 crc kubenswrapper[4726]: I1004 04:03:51.743197 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerID="7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb" exitCode=0 Oct 04 04:03:51 crc kubenswrapper[4726]: I1004 04:03:51.743318 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerDied","Data":"7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb"} Oct 04 04:03:51 crc kubenswrapper[4726]: I1004 04:03:51.743776 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerStarted","Data":"324a30ecf3b4be1667196022de6992ccb11b383614d5e197450a49d29864fb32"} Oct 04 04:03:52 crc kubenswrapper[4726]: I1004 04:03:52.768413 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerStarted","Data":"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74"} Oct 04 04:03:53 crc kubenswrapper[4726]: I1004 04:03:53.786004 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerID="c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74" exitCode=0 Oct 04 04:03:53 crc kubenswrapper[4726]: I1004 04:03:53.786071 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerDied","Data":"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74"} Oct 04 04:03:54 crc kubenswrapper[4726]: I1004 04:03:54.798430 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerStarted","Data":"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c"} Oct 04 04:03:54 crc kubenswrapper[4726]: I1004 04:03:54.827431 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vpqb5" podStartSLOduration=3.38184767 podStartE2EDuration="5.827400741s" podCreationTimestamp="2025-10-04 04:03:49 +0000 UTC" firstStartedPulling="2025-10-04 04:03:51.746338948 +0000 UTC m=+1405.920962201" lastFinishedPulling="2025-10-04 04:03:54.191892019 +0000 UTC m=+1408.366515272" observedRunningTime="2025-10-04 04:03:54.8176206 +0000 UTC m=+1408.992243823" watchObservedRunningTime="2025-10-04 04:03:54.827400741 +0000 UTC m=+1409.002023954" Oct 04 04:03:55 crc kubenswrapper[4726]: I1004 04:03:55.699920 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:03:56 crc kubenswrapper[4726]: I1004 04:03:56.677448 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.187725 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="rabbitmq" containerID="cri-o://782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a" gracePeriod=604796 Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.282999 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.283057 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.331469 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.594419 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="rabbitmq" containerID="cri-o://6371c43e8fec9ebcb3a9d3818bf3e23c8d70a5559ddcda6ea9235314e6623f77" gracePeriod=604797 Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.918317 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:00 crc kubenswrapper[4726]: I1004 04:04:00.985753 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:04:02 crc kubenswrapper[4726]: I1004 04:04:02.895447 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vpqb5" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="registry-server" containerID="cri-o://50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c" gracePeriod=2 Oct 04 04:04:02 crc kubenswrapper[4726]: I1004 04:04:02.984014 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:02 crc kubenswrapper[4726]: I1004 04:04:02.987271 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.025923 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.106658 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whx9j\" (UniqueName: \"kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.106740 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.106769 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.208584 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.208649 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.208790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whx9j\" (UniqueName: \"kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.209265 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.209350 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.229746 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whx9j\" (UniqueName: \"kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j\") pod \"certified-operators-76m7g\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.378059 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.479730 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.614907 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities\") pod \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.615090 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfm4\" (UniqueName: \"kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4\") pod \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.615187 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content\") pod \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\" (UID: \"fdf7f8c5-c25a-49c3-9628-81bcc798af5e\") " Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.617169 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities" (OuterVolumeSpecName: "utilities") pod "fdf7f8c5-c25a-49c3-9628-81bcc798af5e" (UID: "fdf7f8c5-c25a-49c3-9628-81bcc798af5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.622463 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4" (OuterVolumeSpecName: "kube-api-access-8dfm4") pod "fdf7f8c5-c25a-49c3-9628-81bcc798af5e" (UID: "fdf7f8c5-c25a-49c3-9628-81bcc798af5e"). InnerVolumeSpecName "kube-api-access-8dfm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.672096 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdf7f8c5-c25a-49c3-9628-81bcc798af5e" (UID: "fdf7f8c5-c25a-49c3-9628-81bcc798af5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.717826 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfm4\" (UniqueName: \"kubernetes.io/projected/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-kube-api-access-8dfm4\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.717859 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.717868 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdf7f8c5-c25a-49c3-9628-81bcc798af5e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.844668 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.906528 4726 generic.go:334] "Generic (PLEG): container finished" podID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerID="50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c" exitCode=0 Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.906591 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vpqb5" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.906594 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerDied","Data":"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c"} Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.906703 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vpqb5" event={"ID":"fdf7f8c5-c25a-49c3-9628-81bcc798af5e","Type":"ContainerDied","Data":"324a30ecf3b4be1667196022de6992ccb11b383614d5e197450a49d29864fb32"} Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.906721 4726 scope.go:117] "RemoveContainer" containerID="50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.908641 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerStarted","Data":"cf2986b24cc906ae2d3f5674c0ebddb8d82174213a8f564c0ca799d997c78998"} Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.944575 4726 scope.go:117] "RemoveContainer" containerID="c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74" Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.951293 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.958787 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vpqb5"] Oct 04 04:04:03 crc kubenswrapper[4726]: I1004 04:04:03.988826 4726 scope.go:117] "RemoveContainer" containerID="7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.029581 4726 scope.go:117] "RemoveContainer" containerID="50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c" Oct 04 04:04:04 crc kubenswrapper[4726]: E1004 04:04:04.030014 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c\": container with ID starting with 50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c not found: ID does not exist" containerID="50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.030044 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c"} err="failed to get container status \"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c\": rpc error: code = NotFound desc = could not find container \"50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c\": container with ID starting with 50453c69a5da8dac8f613c340cc1c5143d44658dbf1f016b473c642f23765f2c not found: ID does not exist" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.030064 4726 scope.go:117] "RemoveContainer" containerID="c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74" Oct 04 04:04:04 crc kubenswrapper[4726]: E1004 04:04:04.030583 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74\": container with ID starting with c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74 not found: ID does not exist" containerID="c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.030608 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74"} err="failed to get container status \"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74\": rpc error: code = NotFound desc = could not find container \"c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74\": container with ID starting with c8bf68e1f35a4ff0619eba94028e5f432be1d8a0e01e6b5733445de06afe6c74 not found: ID does not exist" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.030677 4726 scope.go:117] "RemoveContainer" containerID="7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb" Oct 04 04:04:04 crc kubenswrapper[4726]: E1004 04:04:04.031180 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb\": container with ID starting with 7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb not found: ID does not exist" containerID="7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.031268 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb"} err="failed to get container status \"7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb\": rpc error: code = NotFound desc = could not find container \"7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb\": container with ID starting with 7c33ed27795f41140ffc64631edf75270d817bde6658c335aed29d779e819cfb not found: ID does not exist" Oct 04 04:04:04 crc kubenswrapper[4726]: E1004 04:04:04.440090 4726 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.176:56750->38.102.83.176:46161: write tcp 38.102.83.176:56750->38.102.83.176:46161: write: broken pipe Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.521689 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" path="/var/lib/kubelet/pods/fdf7f8c5-c25a-49c3-9628-81bcc798af5e/volumes" Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.929275 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerDied","Data":"3129ada17339d029d525d18384d4d8f8f9fe2cc8f235479a74493c0557716343"} Oct 04 04:04:04 crc kubenswrapper[4726]: I1004 04:04:04.933244 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerID="3129ada17339d029d525d18384d4d8f8f9fe2cc8f235479a74493c0557716343" exitCode=0 Oct 04 04:04:05 crc kubenswrapper[4726]: I1004 04:04:05.956093 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerStarted","Data":"44b779fc20873550d9b8d796acf22578c9ea02d19223571dbc648f911ce5c30c"} Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.491739 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.798879 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.906436 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.981000 4726 generic.go:334] "Generic (PLEG): container finished" podID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerID="6371c43e8fec9ebcb3a9d3818bf3e23c8d70a5559ddcda6ea9235314e6623f77" exitCode=0 Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.981096 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerDied","Data":"6371c43e8fec9ebcb3a9d3818bf3e23c8d70a5559ddcda6ea9235314e6623f77"} Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.983217 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerID="44b779fc20873550d9b8d796acf22578c9ea02d19223571dbc648f911ce5c30c" exitCode=0 Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.983266 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerDied","Data":"44b779fc20873550d9b8d796acf22578c9ea02d19223571dbc648f911ce5c30c"} Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988066 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988133 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwldb\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988167 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988195 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988316 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988363 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988382 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988427 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988490 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.988527 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls\") pod \"e7f9fe7b-1700-4842-bda1-354de76c325c\" (UID: \"e7f9fe7b-1700-4842-bda1-354de76c325c\") " Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.989826 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.989855 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.989956 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.992978 4726 generic.go:334] "Generic (PLEG): container finished" podID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerID="782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a" exitCode=0 Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.993026 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerDied","Data":"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a"} Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.993058 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e7f9fe7b-1700-4842-bda1-354de76c325c","Type":"ContainerDied","Data":"989a8f262ff96d221b13e88780dbd27de380ade6d7dc2374a836190da0bb1306"} Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.993078 4726 scope.go:117] "RemoveContainer" containerID="782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.993131 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:04:06 crc kubenswrapper[4726]: I1004 04:04:06.997363 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb" (OuterVolumeSpecName: "kube-api-access-dwldb") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "kube-api-access-dwldb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.002225 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info" (OuterVolumeSpecName: "pod-info") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.019302 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.020142 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.023257 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.039926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data" (OuterVolumeSpecName: "config-data") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091178 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091205 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e7f9fe7b-1700-4842-bda1-354de76c325c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091228 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091237 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091248 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091257 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwldb\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-kube-api-access-dwldb\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091265 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091273 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091281 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e7f9fe7b-1700-4842-bda1-354de76c325c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.091908 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf" (OuterVolumeSpecName: "server-conf") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.111552 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.155299 4726 scope.go:117] "RemoveContainer" containerID="ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.157631 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.176789 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e7f9fe7b-1700-4842-bda1-354de76c325c" (UID: "e7f9fe7b-1700-4842-bda1-354de76c325c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.194271 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.194300 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e7f9fe7b-1700-4842-bda1-354de76c325c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.194314 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e7f9fe7b-1700-4842-bda1-354de76c325c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.206069 4726 scope.go:117] "RemoveContainer" containerID="782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.207168 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a\": container with ID starting with 782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a not found: ID does not exist" containerID="782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.207208 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a"} err="failed to get container status \"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a\": rpc error: code = NotFound desc = could not find container \"782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a\": container with ID starting with 782d397e567d50a608d6b7080bbfc9866e1321a3275a6c7bff2ccbc82a4a688a not found: ID does not exist" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.207235 4726 scope.go:117] "RemoveContainer" containerID="ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.207587 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98\": container with ID starting with ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98 not found: ID does not exist" containerID="ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.207618 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98"} err="failed to get container status \"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98\": rpc error: code = NotFound desc = could not find container \"ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98\": container with ID starting with ae5ed73b2f436a7a8c482ee28c731ccb4472605a5cc3d34df927fc3a085d5b98 not found: ID does not exist" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.294812 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295033 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295152 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295231 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295235 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295432 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295497 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295577 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295722 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295818 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm2kg\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295966 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.296063 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf\") pod \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\" (UID: \"af5efa33-a191-461b-baf3-9b12f2dbd4aa\") " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.295993 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.296493 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.296886 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.296974 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.297059 4726 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.298700 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info" (OuterVolumeSpecName: "pod-info") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.299206 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.299706 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.300496 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg" (OuterVolumeSpecName: "kube-api-access-zm2kg") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "kube-api-access-zm2kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.301184 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.337435 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data" (OuterVolumeSpecName: "config-data") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.363717 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.396739 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399476 4726 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/af5efa33-a191-461b-baf3-9b12f2dbd4aa-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399552 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399786 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399803 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399819 4726 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/af5efa33-a191-461b-baf3-9b12f2dbd4aa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.399836 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm2kg\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-kube-api-access-zm2kg\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.405545 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf" (OuterVolumeSpecName: "server-conf") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447166 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447660 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="extract-utilities" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447680 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="extract-utilities" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447697 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447705 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447719 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="registry-server" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447727 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="registry-server" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447740 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="extract-content" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447747 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="extract-content" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447762 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="setup-container" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447770 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="setup-container" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447784 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447790 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: E1004 04:04:07.447816 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="setup-container" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.447821 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="setup-container" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.448006 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.448020 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" containerName="rabbitmq" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.448039 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf7f8c5-c25a-49c3-9628-81bcc798af5e" containerName="registry-server" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.449211 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.451922 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452151 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "af5efa33-a191-461b-baf3-9b12f2dbd4aa" (UID: "af5efa33-a191-461b-baf3-9b12f2dbd4aa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452235 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452296 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452306 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452377 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452420 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-p9vjw" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.452684 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.481014 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.487722 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502066 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502494 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502548 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502624 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502650 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502707 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502755 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxjsx\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-kube-api-access-rxjsx\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.502808 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.503029 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.503131 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.503444 4726 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/af5efa33-a191-461b-baf3-9b12f2dbd4aa-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.503467 4726 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/af5efa33-a191-461b-baf3-9b12f2dbd4aa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.503479 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.604969 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605395 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxjsx\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-kube-api-access-rxjsx\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605426 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605460 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605495 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605708 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-config-data\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605926 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.606066 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.606687 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.605633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.609652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.609684 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.609820 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.610240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.611427 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.613529 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.613767 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.614263 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.615634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.637878 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxjsx\" (UniqueName: \"kubernetes.io/projected/4ede7a9f-b13d-4fa4-b5b4-9a5f78826351-kube-api-access-rxjsx\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.647852 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351\") " pod="openstack/rabbitmq-server-0" Oct 04 04:04:07 crc kubenswrapper[4726]: I1004 04:04:07.771134 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.014646 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"af5efa33-a191-461b-baf3-9b12f2dbd4aa","Type":"ContainerDied","Data":"f41135e4aa0ae60de7e78dcecd27e19bdd17457df8141c36c441d27e807386cf"} Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.015045 4726 scope.go:117] "RemoveContainer" containerID="6371c43e8fec9ebcb3a9d3818bf3e23c8d70a5559ddcda6ea9235314e6623f77" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.014692 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.032532 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerStarted","Data":"bc1609c0011e9e723be446ec2feb8c131e592743729b115e0e78493750bfb41a"} Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.059036 4726 scope.go:117] "RemoveContainer" containerID="d3041b93f177ab188fb6aef46dddca6ee40048662fe2bdc6cdda20d8866fde54" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.102574 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.113225 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.126494 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-76m7g" podStartSLOduration=3.653813988 podStartE2EDuration="6.126477726s" podCreationTimestamp="2025-10-04 04:04:02 +0000 UTC" firstStartedPulling="2025-10-04 04:04:04.932413944 +0000 UTC m=+1419.107037207" lastFinishedPulling="2025-10-04 04:04:07.405077742 +0000 UTC m=+1421.579700945" observedRunningTime="2025-10-04 04:04:08.073469387 +0000 UTC m=+1422.248092610" watchObservedRunningTime="2025-10-04 04:04:08.126477726 +0000 UTC m=+1422.301100939" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.127243 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.129147 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134280 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134539 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134648 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134807 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2zxzd" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134892 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.134917 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.135644 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.155232 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231132 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231229 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231257 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231284 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231301 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231340 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231359 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sntbq\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-kube-api-access-sntbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231390 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231404 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231427 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.231445 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.276688 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: W1004 04:04:08.280266 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ede7a9f_b13d_4fa4_b5b4_9a5f78826351.slice/crio-e9107f16c0fddd82e245ab9740295467a99dd708cbf63e71c80d96c9fe2a4f67 WatchSource:0}: Error finding container e9107f16c0fddd82e245ab9740295467a99dd708cbf63e71c80d96c9fe2a4f67: Status 404 returned error can't find the container with id e9107f16c0fddd82e245ab9740295467a99dd708cbf63e71c80d96c9fe2a4f67 Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334142 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334472 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334502 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334579 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334629 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334652 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334677 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334694 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334728 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sntbq\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-kube-api-access-sntbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.334903 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.335940 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.336051 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.336514 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.337178 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.340634 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.340867 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.341653 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.342321 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.343687 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.353936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sntbq\" (UniqueName: \"kubernetes.io/projected/179e2c3b-cbc2-4417-9228-1b5eea2a1ced-kube-api-access-sntbq\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.380889 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"179e2c3b-cbc2-4417-9228-1b5eea2a1ced\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.448224 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.516796 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af5efa33-a191-461b-baf3-9b12f2dbd4aa" path="/var/lib/kubelet/pods/af5efa33-a191-461b-baf3-9b12f2dbd4aa/volumes" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.517618 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f9fe7b-1700-4842-bda1-354de76c325c" path="/var/lib/kubelet/pods/e7f9fe7b-1700-4842-bda1-354de76c325c/volumes" Oct 04 04:04:08 crc kubenswrapper[4726]: I1004 04:04:08.902580 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:04:08 crc kubenswrapper[4726]: W1004 04:04:08.903467 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod179e2c3b_cbc2_4417_9228_1b5eea2a1ced.slice/crio-48064dbc84cdb4c4b66c339fc0309a9fd26f14fb35bd742f33684421573426e8 WatchSource:0}: Error finding container 48064dbc84cdb4c4b66c339fc0309a9fd26f14fb35bd742f33684421573426e8: Status 404 returned error can't find the container with id 48064dbc84cdb4c4b66c339fc0309a9fd26f14fb35bd742f33684421573426e8 Oct 04 04:04:09 crc kubenswrapper[4726]: I1004 04:04:09.049287 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"179e2c3b-cbc2-4417-9228-1b5eea2a1ced","Type":"ContainerStarted","Data":"48064dbc84cdb4c4b66c339fc0309a9fd26f14fb35bd742f33684421573426e8"} Oct 04 04:04:09 crc kubenswrapper[4726]: I1004 04:04:09.052994 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351","Type":"ContainerStarted","Data":"e9107f16c0fddd82e245ab9740295467a99dd708cbf63e71c80d96c9fe2a4f67"} Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.037125 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.039905 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.041436 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.051025 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.069709 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351","Type":"ContainerStarted","Data":"d33db6becec1ef43773971256af52edc6b746ea8170892ef7400485aac54f539"} Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170033 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170094 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170307 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170556 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170751 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.170822 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgmmx\" (UniqueName: \"kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.273078 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.273266 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.273440 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.274780 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.274952 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.275202 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.275379 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgmmx\" (UniqueName: \"kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.275475 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.276691 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.276853 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.278319 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.278705 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.280325 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.301989 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgmmx\" (UniqueName: \"kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx\") pod \"dnsmasq-dns-5576978c7c-ggjnp\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.361557 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:10 crc kubenswrapper[4726]: I1004 04:04:10.906492 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:10 crc kubenswrapper[4726]: W1004 04:04:10.913884 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a618393_b8e7_48f1_8db3_4aead5e85ce9.slice/crio-571a2be6602bd74e5543c9208af993499485d6a4951e7535496255df233277ac WatchSource:0}: Error finding container 571a2be6602bd74e5543c9208af993499485d6a4951e7535496255df233277ac: Status 404 returned error can't find the container with id 571a2be6602bd74e5543c9208af993499485d6a4951e7535496255df233277ac Oct 04 04:04:11 crc kubenswrapper[4726]: I1004 04:04:11.082310 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"179e2c3b-cbc2-4417-9228-1b5eea2a1ced","Type":"ContainerStarted","Data":"5428f006bf1d0e7eee56aa7a5740201492a5c0de899ac4fbac9a966aa6dadea1"} Oct 04 04:04:11 crc kubenswrapper[4726]: I1004 04:04:11.084526 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" event={"ID":"7a618393-b8e7-48f1-8db3-4aead5e85ce9","Type":"ContainerStarted","Data":"571a2be6602bd74e5543c9208af993499485d6a4951e7535496255df233277ac"} Oct 04 04:04:12 crc kubenswrapper[4726]: I1004 04:04:12.096473 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerID="e10f4a82d12d67b7c936039ff78aecf2cd5d73eb05c4e4ff9dc861deb55430f7" exitCode=0 Oct 04 04:04:12 crc kubenswrapper[4726]: I1004 04:04:12.098203 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" event={"ID":"7a618393-b8e7-48f1-8db3-4aead5e85ce9","Type":"ContainerDied","Data":"e10f4a82d12d67b7c936039ff78aecf2cd5d73eb05c4e4ff9dc861deb55430f7"} Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.114337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" event={"ID":"7a618393-b8e7-48f1-8db3-4aead5e85ce9","Type":"ContainerStarted","Data":"7076b22906650e4a81f52fe2d3d65c509175044105ed5b86ad938d47ff5340c9"} Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.116635 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.378608 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.378711 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.465986 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:13 crc kubenswrapper[4726]: I1004 04:04:13.505801 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" podStartSLOduration=3.505771479 podStartE2EDuration="3.505771479s" podCreationTimestamp="2025-10-04 04:04:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:04:13.149970554 +0000 UTC m=+1427.324593807" watchObservedRunningTime="2025-10-04 04:04:13.505771479 +0000 UTC m=+1427.680394732" Oct 04 04:04:14 crc kubenswrapper[4726]: I1004 04:04:14.216361 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:14 crc kubenswrapper[4726]: I1004 04:04:14.294535 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:16 crc kubenswrapper[4726]: I1004 04:04:16.164816 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-76m7g" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="registry-server" containerID="cri-o://bc1609c0011e9e723be446ec2feb8c131e592743729b115e0e78493750bfb41a" gracePeriod=2 Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.179271 4726 generic.go:334] "Generic (PLEG): container finished" podID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerID="bc1609c0011e9e723be446ec2feb8c131e592743729b115e0e78493750bfb41a" exitCode=0 Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.179370 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerDied","Data":"bc1609c0011e9e723be446ec2feb8c131e592743729b115e0e78493750bfb41a"} Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.290078 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.385636 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whx9j\" (UniqueName: \"kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j\") pod \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.386233 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities\") pod \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.386408 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content\") pod \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\" (UID: \"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26\") " Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.387516 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities" (OuterVolumeSpecName: "utilities") pod "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" (UID: "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.395089 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j" (OuterVolumeSpecName: "kube-api-access-whx9j") pod "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" (UID: "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26"). InnerVolumeSpecName "kube-api-access-whx9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.450995 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" (UID: "ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.488787 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whx9j\" (UniqueName: \"kubernetes.io/projected/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-kube-api-access-whx9j\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.488832 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:17 crc kubenswrapper[4726]: I1004 04:04:17.488846 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.201005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-76m7g" event={"ID":"ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26","Type":"ContainerDied","Data":"cf2986b24cc906ae2d3f5674c0ebddb8d82174213a8f564c0ca799d997c78998"} Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.201064 4726 scope.go:117] "RemoveContainer" containerID="bc1609c0011e9e723be446ec2feb8c131e592743729b115e0e78493750bfb41a" Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.201077 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-76m7g" Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.253937 4726 scope.go:117] "RemoveContainer" containerID="44b779fc20873550d9b8d796acf22578c9ea02d19223571dbc648f911ce5c30c" Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.271502 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.284680 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-76m7g"] Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.308403 4726 scope.go:117] "RemoveContainer" containerID="3129ada17339d029d525d18384d4d8f8f9fe2cc8f235479a74493c0557716343" Oct 04 04:04:18 crc kubenswrapper[4726]: I1004 04:04:18.518640 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" path="/var/lib/kubelet/pods/ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26/volumes" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.363676 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.440596 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.440877 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="dnsmasq-dns" containerID="cri-o://02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821" gracePeriod=10 Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.636084 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-snnnq"] Oct 04 04:04:20 crc kubenswrapper[4726]: E1004 04:04:20.637120 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="extract-content" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.637481 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="extract-content" Oct 04 04:04:20 crc kubenswrapper[4726]: E1004 04:04:20.637544 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="extract-utilities" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.637553 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="extract-utilities" Oct 04 04:04:20 crc kubenswrapper[4726]: E1004 04:04:20.637568 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="registry-server" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.637575 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="registry-server" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.637808 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac7ceba3-ebdf-40e7-bb31-9eb9af3d3a26" containerName="registry-server" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.639403 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.667768 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.667823 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.667820 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-snnnq"] Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.667947 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.668054 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-config\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.668105 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.668218 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4rn\" (UniqueName: \"kubernetes.io/projected/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-kube-api-access-gb4rn\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.668478 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781609 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4rn\" (UniqueName: \"kubernetes.io/projected/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-kube-api-access-gb4rn\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781755 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781819 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781845 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781892 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781923 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-config\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.781960 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.782929 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.783724 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.784117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.784370 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.784673 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-config\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.785155 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.822834 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4rn\" (UniqueName: \"kubernetes.io/projected/0b74ac9e-0f00-4a57-a05c-a5a155fd51f9-kube-api-access-gb4rn\") pod \"dnsmasq-dns-8c6f6df99-snnnq\" (UID: \"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9\") " pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.970655 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:20 crc kubenswrapper[4726]: I1004 04:04:20.989653 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088494 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsnk6\" (UniqueName: \"kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088537 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088638 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088659 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.088771 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb\") pod \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\" (UID: \"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9\") " Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.098383 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6" (OuterVolumeSpecName: "kube-api-access-zsnk6") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "kube-api-access-zsnk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.149092 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.164185 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config" (OuterVolumeSpecName: "config") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.169689 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.172215 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.187871 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" (UID: "d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190288 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190350 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190363 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190394 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190404 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsnk6\" (UniqueName: \"kubernetes.io/projected/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-kube-api-access-zsnk6\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.190413 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.258660 4726 generic.go:334] "Generic (PLEG): container finished" podID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerID="02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821" exitCode=0 Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.258719 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" event={"ID":"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9","Type":"ContainerDied","Data":"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821"} Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.258752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" event={"ID":"d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9","Type":"ContainerDied","Data":"f44cffa2404e3a34fcfa8ffc472ad52a97f600bba8fe3fc229368bd7af634fa1"} Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.258775 4726 scope.go:117] "RemoveContainer" containerID="02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.258786 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.280586 4726 scope.go:117] "RemoveContainer" containerID="de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.303334 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.310934 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-spvr7"] Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.324341 4726 scope.go:117] "RemoveContainer" containerID="02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821" Oct 04 04:04:21 crc kubenswrapper[4726]: E1004 04:04:21.324869 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821\": container with ID starting with 02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821 not found: ID does not exist" containerID="02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.324906 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821"} err="failed to get container status \"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821\": rpc error: code = NotFound desc = could not find container \"02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821\": container with ID starting with 02bb12a0420e6a474b3fd5fadc63aa10a8022097cbcc2855f73dce24a5c9b821 not found: ID does not exist" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.324931 4726 scope.go:117] "RemoveContainer" containerID="de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e" Oct 04 04:04:21 crc kubenswrapper[4726]: E1004 04:04:21.325193 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e\": container with ID starting with de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e not found: ID does not exist" containerID="de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.325219 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e"} err="failed to get container status \"de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e\": rpc error: code = NotFound desc = could not find container \"de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e\": container with ID starting with de220ddbf784d565b2cf611d47cf30df236c255ef614d4ece8c7ad37a158b73e not found: ID does not exist" Oct 04 04:04:21 crc kubenswrapper[4726]: I1004 04:04:21.427621 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-snnnq"] Oct 04 04:04:21 crc kubenswrapper[4726]: W1004 04:04:21.434812 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b74ac9e_0f00_4a57_a05c_a5a155fd51f9.slice/crio-e7b963da3ea17bdb537485c66c01b4184e14e618e9b485df7d9f8b7f77e8f119 WatchSource:0}: Error finding container e7b963da3ea17bdb537485c66c01b4184e14e618e9b485df7d9f8b7f77e8f119: Status 404 returned error can't find the container with id e7b963da3ea17bdb537485c66c01b4184e14e618e9b485df7d9f8b7f77e8f119 Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.008451 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:22 crc kubenswrapper[4726]: E1004 04:04:22.008885 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="init" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.008905 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="init" Oct 04 04:04:22 crc kubenswrapper[4726]: E1004 04:04:22.008933 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="dnsmasq-dns" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.008940 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="dnsmasq-dns" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.009130 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="dnsmasq-dns" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.010551 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.022414 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.110552 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.110887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g62cd\" (UniqueName: \"kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.110942 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.212397 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g62cd\" (UniqueName: \"kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.212499 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.212635 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.213271 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.213323 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.239008 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g62cd\" (UniqueName: \"kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd\") pod \"redhat-marketplace-vflvz\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.268269 4726 generic.go:334] "Generic (PLEG): container finished" podID="0b74ac9e-0f00-4a57-a05c-a5a155fd51f9" containerID="a7aa0414aed2716e60fe8a4a839a7bdeefcadf475fad5145617cf286308075d8" exitCode=0 Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.268340 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" event={"ID":"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9","Type":"ContainerDied","Data":"a7aa0414aed2716e60fe8a4a839a7bdeefcadf475fad5145617cf286308075d8"} Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.268367 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" event={"ID":"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9","Type":"ContainerStarted","Data":"e7b963da3ea17bdb537485c66c01b4184e14e618e9b485df7d9f8b7f77e8f119"} Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.333710 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.533203 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" path="/var/lib/kubelet/pods/d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9/volumes" Oct 04 04:04:22 crc kubenswrapper[4726]: I1004 04:04:22.804508 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.282454 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" event={"ID":"0b74ac9e-0f00-4a57-a05c-a5a155fd51f9","Type":"ContainerStarted","Data":"5cb9372b8b7e7e055a318e31ce1a4433edc71a963188d735f991fc2153d35bd9"} Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.282850 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.284675 4726 generic.go:334] "Generic (PLEG): container finished" podID="363c2997-5692-4c60-8e8a-15bfb1019248" containerID="4d12216a7452549d7717e99df15b5b3f840797910fddcb7ad8a4f192b47eaf8c" exitCode=0 Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.284713 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerDied","Data":"4d12216a7452549d7717e99df15b5b3f840797910fddcb7ad8a4f192b47eaf8c"} Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.284736 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerStarted","Data":"bafbfea93362c7c3a197f0b80d5579e0626268afa133d05560c2af070b2fcd16"} Oct 04 04:04:23 crc kubenswrapper[4726]: I1004 04:04:23.310996 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" podStartSLOduration=3.310972033 podStartE2EDuration="3.310972033s" podCreationTimestamp="2025-10-04 04:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:04:23.303401088 +0000 UTC m=+1437.478024321" watchObservedRunningTime="2025-10-04 04:04:23.310972033 +0000 UTC m=+1437.485595246" Oct 04 04:04:24 crc kubenswrapper[4726]: I1004 04:04:24.307456 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerStarted","Data":"cb09f557a83e3485f269200b4d32cf59e40c8f4027046324fa4e567da8894ba4"} Oct 04 04:04:25 crc kubenswrapper[4726]: I1004 04:04:25.322467 4726 generic.go:334] "Generic (PLEG): container finished" podID="363c2997-5692-4c60-8e8a-15bfb1019248" containerID="cb09f557a83e3485f269200b4d32cf59e40c8f4027046324fa4e567da8894ba4" exitCode=0 Oct 04 04:04:25 crc kubenswrapper[4726]: I1004 04:04:25.322513 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerDied","Data":"cb09f557a83e3485f269200b4d32cf59e40c8f4027046324fa4e567da8894ba4"} Oct 04 04:04:25 crc kubenswrapper[4726]: I1004 04:04:25.736364 4726 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6c5df9-spvr7" podUID="d7ce6e1e-e673-41b5-b3ae-47cc0c8cc2f9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.200:5353: i/o timeout" Oct 04 04:04:26 crc kubenswrapper[4726]: I1004 04:04:26.343680 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerStarted","Data":"83100978dcd55367b1b844028edab28cf095b5d26392d709af8a1eb48d39876b"} Oct 04 04:04:26 crc kubenswrapper[4726]: I1004 04:04:26.369094 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vflvz" podStartSLOduration=2.801740592 podStartE2EDuration="5.369070072s" podCreationTimestamp="2025-10-04 04:04:21 +0000 UTC" firstStartedPulling="2025-10-04 04:04:23.287646681 +0000 UTC m=+1437.462269904" lastFinishedPulling="2025-10-04 04:04:25.854976151 +0000 UTC m=+1440.029599384" observedRunningTime="2025-10-04 04:04:26.366729682 +0000 UTC m=+1440.541352885" watchObservedRunningTime="2025-10-04 04:04:26.369070072 +0000 UTC m=+1440.543693295" Oct 04 04:04:30 crc kubenswrapper[4726]: I1004 04:04:30.972620 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-snnnq" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.103016 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.103329 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="dnsmasq-dns" containerID="cri-o://7076b22906650e4a81f52fe2d3d65c509175044105ed5b86ad938d47ff5340c9" gracePeriod=10 Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.409915 4726 generic.go:334] "Generic (PLEG): container finished" podID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerID="7076b22906650e4a81f52fe2d3d65c509175044105ed5b86ad938d47ff5340c9" exitCode=0 Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.410024 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" event={"ID":"7a618393-b8e7-48f1-8db3-4aead5e85ce9","Type":"ContainerDied","Data":"7076b22906650e4a81f52fe2d3d65c509175044105ed5b86ad938d47ff5340c9"} Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.586442 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.625803 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626147 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626330 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626416 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626515 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626589 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgmmx\" (UniqueName: \"kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.626707 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb\") pod \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\" (UID: \"7a618393-b8e7-48f1-8db3-4aead5e85ce9\") " Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.632155 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx" (OuterVolumeSpecName: "kube-api-access-dgmmx") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "kube-api-access-dgmmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.673900 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.674811 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.677783 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.683001 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.683722 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.697744 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config" (OuterVolumeSpecName: "config") pod "7a618393-b8e7-48f1-8db3-4aead5e85ce9" (UID: "7a618393-b8e7-48f1-8db3-4aead5e85ce9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729170 4726 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729218 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgmmx\" (UniqueName: \"kubernetes.io/projected/7a618393-b8e7-48f1-8db3-4aead5e85ce9-kube-api-access-dgmmx\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729233 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729248 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729261 4726 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729273 4726 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:31 crc kubenswrapper[4726]: I1004 04:04:31.729284 4726 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a618393-b8e7-48f1-8db3-4aead5e85ce9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.335574 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.336266 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.410495 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.424514 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" event={"ID":"7a618393-b8e7-48f1-8db3-4aead5e85ce9","Type":"ContainerDied","Data":"571a2be6602bd74e5543c9208af993499485d6a4951e7535496255df233277ac"} Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.424546 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-ggjnp" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.424604 4726 scope.go:117] "RemoveContainer" containerID="7076b22906650e4a81f52fe2d3d65c509175044105ed5b86ad938d47ff5340c9" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.464015 4726 scope.go:117] "RemoveContainer" containerID="e10f4a82d12d67b7c936039ff78aecf2cd5d73eb05c4e4ff9dc861deb55430f7" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.477005 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.487223 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-ggjnp"] Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.502224 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.518231 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" path="/var/lib/kubelet/pods/7a618393-b8e7-48f1-8db3-4aead5e85ce9/volumes" Oct 04 04:04:32 crc kubenswrapper[4726]: I1004 04:04:32.657163 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:34 crc kubenswrapper[4726]: I1004 04:04:34.452132 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vflvz" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="registry-server" containerID="cri-o://83100978dcd55367b1b844028edab28cf095b5d26392d709af8a1eb48d39876b" gracePeriod=2 Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.462904 4726 generic.go:334] "Generic (PLEG): container finished" podID="363c2997-5692-4c60-8e8a-15bfb1019248" containerID="83100978dcd55367b1b844028edab28cf095b5d26392d709af8a1eb48d39876b" exitCode=0 Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.462953 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerDied","Data":"83100978dcd55367b1b844028edab28cf095b5d26392d709af8a1eb48d39876b"} Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.462977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vflvz" event={"ID":"363c2997-5692-4c60-8e8a-15bfb1019248","Type":"ContainerDied","Data":"bafbfea93362c7c3a197f0b80d5579e0626268afa133d05560c2af070b2fcd16"} Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.462988 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bafbfea93362c7c3a197f0b80d5579e0626268afa133d05560c2af070b2fcd16" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.467224 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.528337 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content\") pod \"363c2997-5692-4c60-8e8a-15bfb1019248\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.528950 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities\") pod \"363c2997-5692-4c60-8e8a-15bfb1019248\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.529141 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g62cd\" (UniqueName: \"kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd\") pod \"363c2997-5692-4c60-8e8a-15bfb1019248\" (UID: \"363c2997-5692-4c60-8e8a-15bfb1019248\") " Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.531139 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities" (OuterVolumeSpecName: "utilities") pod "363c2997-5692-4c60-8e8a-15bfb1019248" (UID: "363c2997-5692-4c60-8e8a-15bfb1019248"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.540356 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd" (OuterVolumeSpecName: "kube-api-access-g62cd") pod "363c2997-5692-4c60-8e8a-15bfb1019248" (UID: "363c2997-5692-4c60-8e8a-15bfb1019248"). InnerVolumeSpecName "kube-api-access-g62cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.545186 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "363c2997-5692-4c60-8e8a-15bfb1019248" (UID: "363c2997-5692-4c60-8e8a-15bfb1019248"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.646707 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.646747 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363c2997-5692-4c60-8e8a-15bfb1019248-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:35 crc kubenswrapper[4726]: I1004 04:04:35.646757 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g62cd\" (UniqueName: \"kubernetes.io/projected/363c2997-5692-4c60-8e8a-15bfb1019248-kube-api-access-g62cd\") on node \"crc\" DevicePath \"\"" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.473540 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vflvz" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476163 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:04:36 crc kubenswrapper[4726]: E1004 04:04:36.476669 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="registry-server" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476690 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="registry-server" Oct 04 04:04:36 crc kubenswrapper[4726]: E1004 04:04:36.476705 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="init" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476713 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="init" Oct 04 04:04:36 crc kubenswrapper[4726]: E1004 04:04:36.476730 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="extract-content" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476739 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="extract-content" Oct 04 04:04:36 crc kubenswrapper[4726]: E1004 04:04:36.476773 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="extract-utilities" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476781 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="extract-utilities" Oct 04 04:04:36 crc kubenswrapper[4726]: E1004 04:04:36.476795 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="dnsmasq-dns" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.476802 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="dnsmasq-dns" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.477043 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a618393-b8e7-48f1-8db3-4aead5e85ce9" containerName="dnsmasq-dns" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.477066 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" containerName="registry-server" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.478691 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.489039 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.533019 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.541093 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vflvz"] Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.565459 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.565508 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnbjm\" (UniqueName: \"kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.565684 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.667567 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.667621 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnbjm\" (UniqueName: \"kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.667665 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.668008 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.668065 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.703331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnbjm\" (UniqueName: \"kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm\") pod \"redhat-operators-bvzqf\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:36 crc kubenswrapper[4726]: I1004 04:04:36.815817 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:37 crc kubenswrapper[4726]: I1004 04:04:37.289356 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:04:37 crc kubenswrapper[4726]: W1004 04:04:37.298649 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb04e4735_1654_41cf_99c5_1062765ff371.slice/crio-feab49f09cca8c5e0ff2089df10c5f7dbb1bd3f2c8bb6782b3f8040a16ae51bc WatchSource:0}: Error finding container feab49f09cca8c5e0ff2089df10c5f7dbb1bd3f2c8bb6782b3f8040a16ae51bc: Status 404 returned error can't find the container with id feab49f09cca8c5e0ff2089df10c5f7dbb1bd3f2c8bb6782b3f8040a16ae51bc Oct 04 04:04:37 crc kubenswrapper[4726]: I1004 04:04:37.483808 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerStarted","Data":"feab49f09cca8c5e0ff2089df10c5f7dbb1bd3f2c8bb6782b3f8040a16ae51bc"} Oct 04 04:04:38 crc kubenswrapper[4726]: I1004 04:04:38.496174 4726 generic.go:334] "Generic (PLEG): container finished" podID="b04e4735-1654-41cf-99c5-1062765ff371" containerID="468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4" exitCode=0 Oct 04 04:04:38 crc kubenswrapper[4726]: I1004 04:04:38.496275 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerDied","Data":"468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4"} Oct 04 04:04:38 crc kubenswrapper[4726]: I1004 04:04:38.525802 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363c2997-5692-4c60-8e8a-15bfb1019248" path="/var/lib/kubelet/pods/363c2997-5692-4c60-8e8a-15bfb1019248/volumes" Oct 04 04:04:40 crc kubenswrapper[4726]: I1004 04:04:40.526054 4726 generic.go:334] "Generic (PLEG): container finished" podID="b04e4735-1654-41cf-99c5-1062765ff371" containerID="828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c" exitCode=0 Oct 04 04:04:40 crc kubenswrapper[4726]: I1004 04:04:40.529132 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerDied","Data":"828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c"} Oct 04 04:04:41 crc kubenswrapper[4726]: I1004 04:04:41.536885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerStarted","Data":"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767"} Oct 04 04:04:41 crc kubenswrapper[4726]: I1004 04:04:41.574254 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bvzqf" podStartSLOduration=2.795876807 podStartE2EDuration="5.574219254s" podCreationTimestamp="2025-10-04 04:04:36 +0000 UTC" firstStartedPulling="2025-10-04 04:04:38.498698705 +0000 UTC m=+1452.673321928" lastFinishedPulling="2025-10-04 04:04:41.277041152 +0000 UTC m=+1455.451664375" observedRunningTime="2025-10-04 04:04:41.560046578 +0000 UTC m=+1455.734669791" watchObservedRunningTime="2025-10-04 04:04:41.574219254 +0000 UTC m=+1455.748842527" Oct 04 04:04:42 crc kubenswrapper[4726]: I1004 04:04:42.547949 4726 generic.go:334] "Generic (PLEG): container finished" podID="4ede7a9f-b13d-4fa4-b5b4-9a5f78826351" containerID="d33db6becec1ef43773971256af52edc6b746ea8170892ef7400485aac54f539" exitCode=0 Oct 04 04:04:42 crc kubenswrapper[4726]: I1004 04:04:42.548006 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351","Type":"ContainerDied","Data":"d33db6becec1ef43773971256af52edc6b746ea8170892ef7400485aac54f539"} Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.535356 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2"] Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.539221 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.543014 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.569437 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.569498 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.569590 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.571661 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2"] Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.582909 4726 generic.go:334] "Generic (PLEG): container finished" podID="179e2c3b-cbc2-4417-9228-1b5eea2a1ced" containerID="5428f006bf1d0e7eee56aa7a5740201492a5c0de899ac4fbac9a966aa6dadea1" exitCode=0 Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.582986 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"179e2c3b-cbc2-4417-9228-1b5eea2a1ced","Type":"ContainerDied","Data":"5428f006bf1d0e7eee56aa7a5740201492a5c0de899ac4fbac9a966aa6dadea1"} Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.590398 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4ede7a9f-b13d-4fa4-b5b4-9a5f78826351","Type":"ContainerStarted","Data":"0787187224a72bf8baa842018058bd2ccf0373f7897a2dd66a1de3c73a6bbc13"} Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.590671 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.604083 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtjj\" (UniqueName: \"kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.604148 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.604305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.604353 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.705813 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtjj\" (UniqueName: \"kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.705861 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.705989 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.706041 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.710642 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.710888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.718980 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.721205 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtjj\" (UniqueName: \"kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:43 crc kubenswrapper[4726]: I1004 04:04:43.891852 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.536613 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.536593011 podStartE2EDuration="37.536593011s" podCreationTimestamp="2025-10-04 04:04:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:04:43.65181723 +0000 UTC m=+1457.826440433" watchObservedRunningTime="2025-10-04 04:04:44.536593011 +0000 UTC m=+1458.711216224" Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.542068 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2"] Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.612232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" event={"ID":"db0467b4-3329-4278-b094-02b70e240f01","Type":"ContainerStarted","Data":"c3eec13038a971ebcbeb0395f97bc74a552b8f9ea7d64ecf0ee137475737dfc2"} Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.615122 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"179e2c3b-cbc2-4417-9228-1b5eea2a1ced","Type":"ContainerStarted","Data":"29c5b9fa03e9900c697d215c8ff05fa169211f04b7b04b5812cf9637636287a2"} Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.615499 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:04:44 crc kubenswrapper[4726]: I1004 04:04:44.643018 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.643001708 podStartE2EDuration="36.643001708s" podCreationTimestamp="2025-10-04 04:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:04:44.637833725 +0000 UTC m=+1458.812456938" watchObservedRunningTime="2025-10-04 04:04:44.643001708 +0000 UTC m=+1458.817624921" Oct 04 04:04:46 crc kubenswrapper[4726]: I1004 04:04:46.816517 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:46 crc kubenswrapper[4726]: I1004 04:04:46.817734 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:04:47 crc kubenswrapper[4726]: I1004 04:04:47.860501 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bvzqf" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" probeResult="failure" output=< Oct 04 04:04:47 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:04:47 crc kubenswrapper[4726]: > Oct 04 04:04:55 crc kubenswrapper[4726]: I1004 04:04:55.764254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" event={"ID":"db0467b4-3329-4278-b094-02b70e240f01","Type":"ContainerStarted","Data":"b0cf45bc559afd1d1b51892ef40ddead00c19584884a8947681add1ff6f6693e"} Oct 04 04:04:55 crc kubenswrapper[4726]: I1004 04:04:55.783814 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" podStartSLOduration=2.389717586 podStartE2EDuration="12.783795413s" podCreationTimestamp="2025-10-04 04:04:43 +0000 UTC" firstStartedPulling="2025-10-04 04:04:44.543173891 +0000 UTC m=+1458.717797114" lastFinishedPulling="2025-10-04 04:04:54.937251738 +0000 UTC m=+1469.111874941" observedRunningTime="2025-10-04 04:04:55.781453102 +0000 UTC m=+1469.956076325" watchObservedRunningTime="2025-10-04 04:04:55.783795413 +0000 UTC m=+1469.958418626" Oct 04 04:04:57 crc kubenswrapper[4726]: I1004 04:04:57.776500 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:04:57 crc kubenswrapper[4726]: I1004 04:04:57.872342 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bvzqf" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" probeResult="failure" output=< Oct 04 04:04:57 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:04:57 crc kubenswrapper[4726]: > Oct 04 04:04:58 crc kubenswrapper[4726]: I1004 04:04:58.451375 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:05:06 crc kubenswrapper[4726]: I1004 04:05:06.876645 4726 generic.go:334] "Generic (PLEG): container finished" podID="db0467b4-3329-4278-b094-02b70e240f01" containerID="b0cf45bc559afd1d1b51892ef40ddead00c19584884a8947681add1ff6f6693e" exitCode=0 Oct 04 04:05:06 crc kubenswrapper[4726]: I1004 04:05:06.876752 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" event={"ID":"db0467b4-3329-4278-b094-02b70e240f01","Type":"ContainerDied","Data":"b0cf45bc559afd1d1b51892ef40ddead00c19584884a8947681add1ff6f6693e"} Oct 04 04:05:07 crc kubenswrapper[4726]: I1004 04:05:07.887531 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bvzqf" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" probeResult="failure" output=< Oct 04 04:05:07 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:05:07 crc kubenswrapper[4726]: > Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.358900 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.422373 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory\") pod \"db0467b4-3329-4278-b094-02b70e240f01\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.422433 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle\") pod \"db0467b4-3329-4278-b094-02b70e240f01\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.422520 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldtjj\" (UniqueName: \"kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj\") pod \"db0467b4-3329-4278-b094-02b70e240f01\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.422568 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key\") pod \"db0467b4-3329-4278-b094-02b70e240f01\" (UID: \"db0467b4-3329-4278-b094-02b70e240f01\") " Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.428563 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "db0467b4-3329-4278-b094-02b70e240f01" (UID: "db0467b4-3329-4278-b094-02b70e240f01"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.428988 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj" (OuterVolumeSpecName: "kube-api-access-ldtjj") pod "db0467b4-3329-4278-b094-02b70e240f01" (UID: "db0467b4-3329-4278-b094-02b70e240f01"). InnerVolumeSpecName "kube-api-access-ldtjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.455503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db0467b4-3329-4278-b094-02b70e240f01" (UID: "db0467b4-3329-4278-b094-02b70e240f01"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.469969 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory" (OuterVolumeSpecName: "inventory") pod "db0467b4-3329-4278-b094-02b70e240f01" (UID: "db0467b4-3329-4278-b094-02b70e240f01"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.525636 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.525820 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.525885 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db0467b4-3329-4278-b094-02b70e240f01-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.525939 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldtjj\" (UniqueName: \"kubernetes.io/projected/db0467b4-3329-4278-b094-02b70e240f01-kube-api-access-ldtjj\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.909950 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" event={"ID":"db0467b4-3329-4278-b094-02b70e240f01","Type":"ContainerDied","Data":"c3eec13038a971ebcbeb0395f97bc74a552b8f9ea7d64ecf0ee137475737dfc2"} Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.910017 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3eec13038a971ebcbeb0395f97bc74a552b8f9ea7d64ecf0ee137475737dfc2" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.910065 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.976678 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds"] Oct 04 04:05:08 crc kubenswrapper[4726]: E1004 04:05:08.977066 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db0467b4-3329-4278-b094-02b70e240f01" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.977080 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="db0467b4-3329-4278-b094-02b70e240f01" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.977278 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="db0467b4-3329-4278-b094-02b70e240f01" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.977831 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.982436 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.982751 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.983158 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.983403 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:05:08 crc kubenswrapper[4726]: I1004 04:05:08.997533 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds"] Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.038305 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.038443 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.038474 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4899\" (UniqueName: \"kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.139939 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.140036 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.140062 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4899\" (UniqueName: \"kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.148441 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.151240 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.172913 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4899\" (UniqueName: \"kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-6mqds\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.298016 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:09 crc kubenswrapper[4726]: I1004 04:05:09.950755 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds"] Oct 04 04:05:10 crc kubenswrapper[4726]: I1004 04:05:10.939513 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" event={"ID":"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad","Type":"ContainerStarted","Data":"7611785e59157ea4fad881994c98ffcacd060d7ec461a8e66dcdc67f37be5303"} Oct 04 04:05:10 crc kubenswrapper[4726]: I1004 04:05:10.940244 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" event={"ID":"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad","Type":"ContainerStarted","Data":"9a531676658c061421134b2430d2f209b2819ff91e929673e9d006061498592f"} Oct 04 04:05:10 crc kubenswrapper[4726]: I1004 04:05:10.964282 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" podStartSLOduration=2.5468226400000002 podStartE2EDuration="2.964256627s" podCreationTimestamp="2025-10-04 04:05:08 +0000 UTC" firstStartedPulling="2025-10-04 04:05:09.960281598 +0000 UTC m=+1484.134904841" lastFinishedPulling="2025-10-04 04:05:10.377715585 +0000 UTC m=+1484.552338828" observedRunningTime="2025-10-04 04:05:10.958916699 +0000 UTC m=+1485.133539942" watchObservedRunningTime="2025-10-04 04:05:10.964256627 +0000 UTC m=+1485.138879850" Oct 04 04:05:13 crc kubenswrapper[4726]: I1004 04:05:13.972641 4726 generic.go:334] "Generic (PLEG): container finished" podID="2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" containerID="7611785e59157ea4fad881994c98ffcacd060d7ec461a8e66dcdc67f37be5303" exitCode=0 Oct 04 04:05:13 crc kubenswrapper[4726]: I1004 04:05:13.972709 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" event={"ID":"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad","Type":"ContainerDied","Data":"7611785e59157ea4fad881994c98ffcacd060d7ec461a8e66dcdc67f37be5303"} Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.457199 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.567021 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory\") pod \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.567088 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key\") pod \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.567211 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4899\" (UniqueName: \"kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899\") pod \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\" (UID: \"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad\") " Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.583347 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899" (OuterVolumeSpecName: "kube-api-access-j4899") pod "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" (UID: "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad"). InnerVolumeSpecName "kube-api-access-j4899". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.606872 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory" (OuterVolumeSpecName: "inventory") pod "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" (UID: "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.615177 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" (UID: "2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.670327 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.670361 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.670375 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4899\" (UniqueName: \"kubernetes.io/projected/2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad-kube-api-access-j4899\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.996158 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" event={"ID":"2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad","Type":"ContainerDied","Data":"9a531676658c061421134b2430d2f209b2819ff91e929673e9d006061498592f"} Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.996200 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a531676658c061421134b2430d2f209b2819ff91e929673e9d006061498592f" Oct 04 04:05:15 crc kubenswrapper[4726]: I1004 04:05:15.996213 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-6mqds" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.077199 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg"] Oct 04 04:05:16 crc kubenswrapper[4726]: E1004 04:05:16.077847 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.077864 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.078064 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.078815 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.083330 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.083455 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.083500 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.083619 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.087932 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg"] Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.178962 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.179251 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92x9g\" (UniqueName: \"kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.179295 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.179329 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.281403 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92x9g\" (UniqueName: \"kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.281469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.281514 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.281568 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.286998 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.287332 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.291279 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.303092 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92x9g\" (UniqueName: \"kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.398749 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:05:16 crc kubenswrapper[4726]: I1004 04:05:16.898383 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:05:17 crc kubenswrapper[4726]: I1004 04:05:17.005343 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg"] Oct 04 04:05:17 crc kubenswrapper[4726]: I1004 04:05:17.009534 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:05:17 crc kubenswrapper[4726]: I1004 04:05:17.014287 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" event={"ID":"aa6f54c4-9900-41b2-a047-6dda8304ff99","Type":"ContainerStarted","Data":"5828620534d38dc787f5a3434c6b641561c42ea9859258ac814cf4f86628a3b9"} Oct 04 04:05:17 crc kubenswrapper[4726]: I1004 04:05:17.139657 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.025650 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bvzqf" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" containerID="cri-o://2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767" gracePeriod=2 Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.028083 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" event={"ID":"aa6f54c4-9900-41b2-a047-6dda8304ff99","Type":"ContainerStarted","Data":"1b9348d2606b84af833cfb43133cd07073570bafb4feebfca88471ff3cc36327"} Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.064578 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" podStartSLOduration=1.5574871209999999 podStartE2EDuration="2.06455367s" podCreationTimestamp="2025-10-04 04:05:16 +0000 UTC" firstStartedPulling="2025-10-04 04:05:17.01065127 +0000 UTC m=+1491.185274493" lastFinishedPulling="2025-10-04 04:05:17.517717829 +0000 UTC m=+1491.692341042" observedRunningTime="2025-10-04 04:05:18.050786289 +0000 UTC m=+1492.225409552" watchObservedRunningTime="2025-10-04 04:05:18.06455367 +0000 UTC m=+1492.239176893" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.501199 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.635164 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities\") pod \"b04e4735-1654-41cf-99c5-1062765ff371\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.635312 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content\") pod \"b04e4735-1654-41cf-99c5-1062765ff371\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.635356 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnbjm\" (UniqueName: \"kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm\") pod \"b04e4735-1654-41cf-99c5-1062765ff371\" (UID: \"b04e4735-1654-41cf-99c5-1062765ff371\") " Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.636363 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities" (OuterVolumeSpecName: "utilities") pod "b04e4735-1654-41cf-99c5-1062765ff371" (UID: "b04e4735-1654-41cf-99c5-1062765ff371"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.638738 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.640476 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm" (OuterVolumeSpecName: "kube-api-access-tnbjm") pod "b04e4735-1654-41cf-99c5-1062765ff371" (UID: "b04e4735-1654-41cf-99c5-1062765ff371"). InnerVolumeSpecName "kube-api-access-tnbjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.704775 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b04e4735-1654-41cf-99c5-1062765ff371" (UID: "b04e4735-1654-41cf-99c5-1062765ff371"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.740362 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b04e4735-1654-41cf-99c5-1062765ff371-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:18 crc kubenswrapper[4726]: I1004 04:05:18.740563 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnbjm\" (UniqueName: \"kubernetes.io/projected/b04e4735-1654-41cf-99c5-1062765ff371-kube-api-access-tnbjm\") on node \"crc\" DevicePath \"\"" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.039085 4726 generic.go:334] "Generic (PLEG): container finished" podID="b04e4735-1654-41cf-99c5-1062765ff371" containerID="2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767" exitCode=0 Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.039232 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerDied","Data":"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767"} Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.040616 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bvzqf" event={"ID":"b04e4735-1654-41cf-99c5-1062765ff371","Type":"ContainerDied","Data":"feab49f09cca8c5e0ff2089df10c5f7dbb1bd3f2c8bb6782b3f8040a16ae51bc"} Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.039273 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bvzqf" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.040685 4726 scope.go:117] "RemoveContainer" containerID="2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.075587 4726 scope.go:117] "RemoveContainer" containerID="828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.095172 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.103973 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bvzqf"] Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.112962 4726 scope.go:117] "RemoveContainer" containerID="468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.150668 4726 scope.go:117] "RemoveContainer" containerID="2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767" Oct 04 04:05:19 crc kubenswrapper[4726]: E1004 04:05:19.151121 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767\": container with ID starting with 2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767 not found: ID does not exist" containerID="2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.151156 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767"} err="failed to get container status \"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767\": rpc error: code = NotFound desc = could not find container \"2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767\": container with ID starting with 2b8d36b9279a363dedd32936ca510f6a78a70820cdcb28309288532126e1d767 not found: ID does not exist" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.151182 4726 scope.go:117] "RemoveContainer" containerID="828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c" Oct 04 04:05:19 crc kubenswrapper[4726]: E1004 04:05:19.151582 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c\": container with ID starting with 828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c not found: ID does not exist" containerID="828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.151647 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c"} err="failed to get container status \"828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c\": rpc error: code = NotFound desc = could not find container \"828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c\": container with ID starting with 828bd4e2f9c2daa9841a612b352ade7e92bd7c67ef39d9c794fe8eba9466325c not found: ID does not exist" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.151690 4726 scope.go:117] "RemoveContainer" containerID="468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4" Oct 04 04:05:19 crc kubenswrapper[4726]: E1004 04:05:19.152258 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4\": container with ID starting with 468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4 not found: ID does not exist" containerID="468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4" Oct 04 04:05:19 crc kubenswrapper[4726]: I1004 04:05:19.152284 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4"} err="failed to get container status \"468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4\": rpc error: code = NotFound desc = could not find container \"468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4\": container with ID starting with 468e1fb5bbd7f6f2311980eabc071aa9c549dbda681a647cf432fd258c4f8bc4 not found: ID does not exist" Oct 04 04:05:20 crc kubenswrapper[4726]: I1004 04:05:20.518223 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b04e4735-1654-41cf-99c5-1062765ff371" path="/var/lib/kubelet/pods/b04e4735-1654-41cf-99c5-1062765ff371/volumes" Oct 04 04:05:34 crc kubenswrapper[4726]: I1004 04:05:34.188655 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:05:34 crc kubenswrapper[4726]: I1004 04:05:34.189602 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:05:34 crc kubenswrapper[4726]: I1004 04:05:34.629231 4726 scope.go:117] "RemoveContainer" containerID="70373a7d6931b053b6b9de4aabd4b8ca0ccbfea6b4803cea33c1113fe542dbb7" Oct 04 04:05:34 crc kubenswrapper[4726]: I1004 04:05:34.674913 4726 scope.go:117] "RemoveContainer" containerID="ae17c06d117565d3fc29d4864e8b4a25223a3a1011f13f7a5aa85c4889cd6d9c" Oct 04 04:06:04 crc kubenswrapper[4726]: I1004 04:06:04.188575 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:06:04 crc kubenswrapper[4726]: I1004 04:06:04.189241 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.188214 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.188817 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.188870 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.189696 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.189750 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" gracePeriod=600 Oct 04 04:06:34 crc kubenswrapper[4726]: E1004 04:06:34.333985 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.790434 4726 scope.go:117] "RemoveContainer" containerID="c788b936e0dbd761637811f8d0c7d0930ef0ed905f894ab01f13cc9839e622ca" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.957888 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" exitCode=0 Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.957939 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9"} Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.957980 4726 scope.go:117] "RemoveContainer" containerID="b2ae2c35dc687883ce23aeb6d29e6ce14b0bb89e46646be08c290a88144230af" Oct 04 04:06:34 crc kubenswrapper[4726]: I1004 04:06:34.958923 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:06:34 crc kubenswrapper[4726]: E1004 04:06:34.959397 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:06:46 crc kubenswrapper[4726]: I1004 04:06:46.510540 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:06:46 crc kubenswrapper[4726]: E1004 04:06:46.511283 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:06:57 crc kubenswrapper[4726]: I1004 04:06:57.503089 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:06:57 crc kubenswrapper[4726]: E1004 04:06:57.503951 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:07:12 crc kubenswrapper[4726]: I1004 04:07:12.502833 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:07:12 crc kubenswrapper[4726]: E1004 04:07:12.504314 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:07:27 crc kubenswrapper[4726]: I1004 04:07:27.502448 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:07:27 crc kubenswrapper[4726]: E1004 04:07:27.503209 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:07:34 crc kubenswrapper[4726]: I1004 04:07:34.897060 4726 scope.go:117] "RemoveContainer" containerID="36c16ca90803489bd416b0a75f9350c121f6743ad6e3bc03208b982406944302" Oct 04 04:07:34 crc kubenswrapper[4726]: I1004 04:07:34.939522 4726 scope.go:117] "RemoveContainer" containerID="cdf750fb434c1778bfd98585f593db00d740410b7f80d2188619ea27d7425865" Oct 04 04:07:42 crc kubenswrapper[4726]: I1004 04:07:42.502393 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:07:42 crc kubenswrapper[4726]: E1004 04:07:42.503433 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:07:53 crc kubenswrapper[4726]: I1004 04:07:53.503635 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:07:53 crc kubenswrapper[4726]: E1004 04:07:53.505279 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:08:07 crc kubenswrapper[4726]: I1004 04:08:07.502037 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:08:07 crc kubenswrapper[4726]: E1004 04:08:07.502923 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:08:21 crc kubenswrapper[4726]: I1004 04:08:21.503356 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:08:21 crc kubenswrapper[4726]: E1004 04:08:21.504271 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:08:33 crc kubenswrapper[4726]: I1004 04:08:33.503010 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:08:33 crc kubenswrapper[4726]: E1004 04:08:33.504147 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:08:34 crc kubenswrapper[4726]: I1004 04:08:34.521776 4726 generic.go:334] "Generic (PLEG): container finished" podID="aa6f54c4-9900-41b2-a047-6dda8304ff99" containerID="1b9348d2606b84af833cfb43133cd07073570bafb4feebfca88471ff3cc36327" exitCode=0 Oct 04 04:08:34 crc kubenswrapper[4726]: I1004 04:08:34.521874 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" event={"ID":"aa6f54c4-9900-41b2-a047-6dda8304ff99","Type":"ContainerDied","Data":"1b9348d2606b84af833cfb43133cd07073570bafb4feebfca88471ff3cc36327"} Oct 04 04:08:35 crc kubenswrapper[4726]: I1004 04:08:35.026291 4726 scope.go:117] "RemoveContainer" containerID="d61e64a1a93a2ac275345ec3b2dc40dbce5172981bf6f137d340956567b48a3e" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.001631 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.016467 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle\") pod \"aa6f54c4-9900-41b2-a047-6dda8304ff99\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.016591 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92x9g\" (UniqueName: \"kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g\") pod \"aa6f54c4-9900-41b2-a047-6dda8304ff99\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.016623 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key\") pod \"aa6f54c4-9900-41b2-a047-6dda8304ff99\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.016674 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory\") pod \"aa6f54c4-9900-41b2-a047-6dda8304ff99\" (UID: \"aa6f54c4-9900-41b2-a047-6dda8304ff99\") " Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.026251 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "aa6f54c4-9900-41b2-a047-6dda8304ff99" (UID: "aa6f54c4-9900-41b2-a047-6dda8304ff99"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.027285 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g" (OuterVolumeSpecName: "kube-api-access-92x9g") pod "aa6f54c4-9900-41b2-a047-6dda8304ff99" (UID: "aa6f54c4-9900-41b2-a047-6dda8304ff99"). InnerVolumeSpecName "kube-api-access-92x9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.061547 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa6f54c4-9900-41b2-a047-6dda8304ff99" (UID: "aa6f54c4-9900-41b2-a047-6dda8304ff99"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.064538 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory" (OuterVolumeSpecName: "inventory") pod "aa6f54c4-9900-41b2-a047-6dda8304ff99" (UID: "aa6f54c4-9900-41b2-a047-6dda8304ff99"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.119162 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.119211 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.119233 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92x9g\" (UniqueName: \"kubernetes.io/projected/aa6f54c4-9900-41b2-a047-6dda8304ff99-kube-api-access-92x9g\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.119252 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa6f54c4-9900-41b2-a047-6dda8304ff99-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.550725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" event={"ID":"aa6f54c4-9900-41b2-a047-6dda8304ff99","Type":"ContainerDied","Data":"5828620534d38dc787f5a3434c6b641561c42ea9859258ac814cf4f86628a3b9"} Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.551171 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5828620534d38dc787f5a3434c6b641561c42ea9859258ac814cf4f86628a3b9" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.550825 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676077 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5"] Oct 04 04:08:36 crc kubenswrapper[4726]: E1004 04:08:36.676620 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676641 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" Oct 04 04:08:36 crc kubenswrapper[4726]: E1004 04:08:36.676679 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="extract-utilities" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676689 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="extract-utilities" Oct 04 04:08:36 crc kubenswrapper[4726]: E1004 04:08:36.676720 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa6f54c4-9900-41b2-a047-6dda8304ff99" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676729 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa6f54c4-9900-41b2-a047-6dda8304ff99" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 04:08:36 crc kubenswrapper[4726]: E1004 04:08:36.676744 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="extract-content" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676752 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="extract-content" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.676978 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa6f54c4-9900-41b2-a047-6dda8304ff99" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.677017 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b04e4735-1654-41cf-99c5-1062765ff371" containerName="registry-server" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.677824 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.681391 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.681403 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.681672 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.682241 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.688174 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5"] Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.833711 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr2hr\" (UniqueName: \"kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.833831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.833920 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.936313 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr2hr\" (UniqueName: \"kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.936690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.937019 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.942977 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.943132 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.959212 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr2hr\" (UniqueName: \"kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:36 crc kubenswrapper[4726]: I1004 04:08:36.998268 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:08:37 crc kubenswrapper[4726]: I1004 04:08:37.347387 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5"] Oct 04 04:08:37 crc kubenswrapper[4726]: W1004 04:08:37.357167 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a4f5e93_bd73_42f9_8282_39e559ad08a7.slice/crio-cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df WatchSource:0}: Error finding container cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df: Status 404 returned error can't find the container with id cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df Oct 04 04:08:37 crc kubenswrapper[4726]: I1004 04:08:37.360489 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:08:37 crc kubenswrapper[4726]: I1004 04:08:37.564210 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" event={"ID":"4a4f5e93-bd73-42f9-8282-39e559ad08a7","Type":"ContainerStarted","Data":"cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df"} Oct 04 04:08:38 crc kubenswrapper[4726]: I1004 04:08:38.580997 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" event={"ID":"4a4f5e93-bd73-42f9-8282-39e559ad08a7","Type":"ContainerStarted","Data":"ca77fc0b7cc20bf5f31370ecabb5883c71d492e92b1864dfb50a15217c3e288e"} Oct 04 04:08:38 crc kubenswrapper[4726]: I1004 04:08:38.613773 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" podStartSLOduration=2.079940196 podStartE2EDuration="2.613748683s" podCreationTimestamp="2025-10-04 04:08:36 +0000 UTC" firstStartedPulling="2025-10-04 04:08:37.36031482 +0000 UTC m=+1691.534938033" lastFinishedPulling="2025-10-04 04:08:37.894123307 +0000 UTC m=+1692.068746520" observedRunningTime="2025-10-04 04:08:38.607042567 +0000 UTC m=+1692.781665790" watchObservedRunningTime="2025-10-04 04:08:38.613748683 +0000 UTC m=+1692.788371906" Oct 04 04:08:48 crc kubenswrapper[4726]: I1004 04:08:48.502801 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:08:48 crc kubenswrapper[4726]: E1004 04:08:48.503676 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:09:02 crc kubenswrapper[4726]: I1004 04:09:02.503418 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:09:02 crc kubenswrapper[4726]: E1004 04:09:02.504681 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.038622 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2jttv"] Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.058035 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-7rbpz"] Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.069983 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jpbkh"] Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.080430 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2jttv"] Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.089670 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-7rbpz"] Oct 04 04:09:05 crc kubenswrapper[4726]: I1004 04:09:05.098168 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jpbkh"] Oct 04 04:09:06 crc kubenswrapper[4726]: I1004 04:09:06.524094 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e76b2f-db99-4f3d-8f36-68df71ff19d1" path="/var/lib/kubelet/pods/25e76b2f-db99-4f3d-8f36-68df71ff19d1/volumes" Oct 04 04:09:06 crc kubenswrapper[4726]: I1004 04:09:06.526069 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2864ee-55ba-4488-8632-cd2e985c7ecc" path="/var/lib/kubelet/pods/3c2864ee-55ba-4488-8632-cd2e985c7ecc/volumes" Oct 04 04:09:06 crc kubenswrapper[4726]: I1004 04:09:06.527335 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6353dfea-176b-4703-8076-afabae252368" path="/var/lib/kubelet/pods/6353dfea-176b-4703-8076-afabae252368/volumes" Oct 04 04:09:14 crc kubenswrapper[4726]: I1004 04:09:14.035366 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-8af8-account-create-gvlrl"] Oct 04 04:09:14 crc kubenswrapper[4726]: I1004 04:09:14.047543 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-8af8-account-create-gvlrl"] Oct 04 04:09:14 crc kubenswrapper[4726]: I1004 04:09:14.521753 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6082cf2a-50dc-4e4e-8908-12b573c30b02" path="/var/lib/kubelet/pods/6082cf2a-50dc-4e4e-8908-12b573c30b02/volumes" Oct 04 04:09:16 crc kubenswrapper[4726]: I1004 04:09:16.516810 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:09:16 crc kubenswrapper[4726]: E1004 04:09:16.517436 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.072303 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lfqcg"] Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.110476 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rlls9"] Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.131662 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xszts"] Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.138786 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lfqcg"] Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.145812 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xszts"] Oct 04 04:09:21 crc kubenswrapper[4726]: I1004 04:09:21.156200 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rlls9"] Oct 04 04:09:22 crc kubenswrapper[4726]: I1004 04:09:22.518230 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a09631ae-f0ad-46af-b441-76ce38840f42" path="/var/lib/kubelet/pods/a09631ae-f0ad-46af-b441-76ce38840f42/volumes" Oct 04 04:09:22 crc kubenswrapper[4726]: I1004 04:09:22.518946 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdee62dd-66dd-42da-840c-520048edfd3e" path="/var/lib/kubelet/pods/cdee62dd-66dd-42da-840c-520048edfd3e/volumes" Oct 04 04:09:22 crc kubenswrapper[4726]: I1004 04:09:22.519620 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e54464f6-5188-4631-a0d3-609ce5b2584c" path="/var/lib/kubelet/pods/e54464f6-5188-4631-a0d3-609ce5b2584c/volumes" Oct 04 04:09:29 crc kubenswrapper[4726]: I1004 04:09:29.503874 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:09:29 crc kubenswrapper[4726]: E1004 04:09:29.505269 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:09:31 crc kubenswrapper[4726]: I1004 04:09:31.053926 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d7ee-account-create-7t464"] Oct 04 04:09:31 crc kubenswrapper[4726]: I1004 04:09:31.069974 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-987d-account-create-qs5z8"] Oct 04 04:09:31 crc kubenswrapper[4726]: I1004 04:09:31.085695 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d7ee-account-create-7t464"] Oct 04 04:09:31 crc kubenswrapper[4726]: I1004 04:09:31.095775 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-987d-account-create-qs5z8"] Oct 04 04:09:32 crc kubenswrapper[4726]: I1004 04:09:32.523754 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb49d8bc-c2a0-41ac-9a64-de6e6f502cec" path="/var/lib/kubelet/pods/cb49d8bc-c2a0-41ac-9a64-de6e6f502cec/volumes" Oct 04 04:09:32 crc kubenswrapper[4726]: I1004 04:09:32.524874 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d99bb218-074c-4ffd-abc8-a40bbbb2c611" path="/var/lib/kubelet/pods/d99bb218-074c-4ffd-abc8-a40bbbb2c611/volumes" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.031755 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-613a-account-create-ccqbc"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.043305 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6ea2-account-create-8tn2m"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.054874 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8344-account-create-95zcb"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.066414 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6ea2-account-create-8tn2m"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.073983 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-613a-account-create-ccqbc"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.089032 4726 scope.go:117] "RemoveContainer" containerID="88ce55993e6aaf7ebed96ae8b671f5efa5f400e6ddc564ee5db20bb74ea7f124" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.090701 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8344-account-create-95zcb"] Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.120052 4726 scope.go:117] "RemoveContainer" containerID="b45e5b6955e72e5c0a7286fef2f30799d7b2f85c701556ce5a6e8e73ee78b41a" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.146629 4726 scope.go:117] "RemoveContainer" containerID="ca6c6ce3b9f0ecfe725e6a57a4de9eb79e470a0e269311612c1cc4afeabb98dd" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.214803 4726 scope.go:117] "RemoveContainer" containerID="5fa8851085155a12f785310916c594cfef41618ba305c4a842fb328eeb4cc89a" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.242148 4726 scope.go:117] "RemoveContainer" containerID="a640bb89d686ebbab0a02db0033f70a8db505b6c165d09ed4fa3753a59c93a82" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.282663 4726 scope.go:117] "RemoveContainer" containerID="d5aa397c8c7499bf062657034bb4a6e79e3a65f258daf3a9748d5e1c5f0afe9d" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.320872 4726 scope.go:117] "RemoveContainer" containerID="c7e1b8a524c53f78071e001f65e63c25b7dfb5133e67957143474537d36ac043" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.341127 4726 scope.go:117] "RemoveContainer" containerID="9d088da5d54b7350c1ee13c6d593be12534721ecf5fd38fb74c8a98fb39f1ac2" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.387220 4726 scope.go:117] "RemoveContainer" containerID="5322e52cbdb2889fbf57cfcc8616ec6c47338bc8d512659a417065eec9a029f1" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.423727 4726 scope.go:117] "RemoveContainer" containerID="7074e87f33d832b0685f9a537bfec335534441db6dbec0ae99d66e3feb93653c" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.443738 4726 scope.go:117] "RemoveContainer" containerID="0e40940d5d00393fab8ee50689d4119e0c30bb714227eba711a8375f701c7c9e" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.462201 4726 scope.go:117] "RemoveContainer" containerID="7d38f7caecb6e1573027ffdabc8a8062865a86c0af63ba56756b2bba80978073" Oct 04 04:09:35 crc kubenswrapper[4726]: I1004 04:09:35.490628 4726 scope.go:117] "RemoveContainer" containerID="0bb9e6de62e1c70b8401b93e237221c882313b93df7a747834e5798b35fb282b" Oct 04 04:09:36 crc kubenswrapper[4726]: I1004 04:09:36.516511 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f850a7-b8df-4af9-a0df-0f78a6f37482" path="/var/lib/kubelet/pods/19f850a7-b8df-4af9-a0df-0f78a6f37482/volumes" Oct 04 04:09:36 crc kubenswrapper[4726]: I1004 04:09:36.517665 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5383a9a4-9771-40c8-b9fb-4fcbe87deb85" path="/var/lib/kubelet/pods/5383a9a4-9771-40c8-b9fb-4fcbe87deb85/volumes" Oct 04 04:09:36 crc kubenswrapper[4726]: I1004 04:09:36.518159 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae06837e-2532-440c-9a0f-65dd2b70b1e7" path="/var/lib/kubelet/pods/ae06837e-2532-440c-9a0f-65dd2b70b1e7/volumes" Oct 04 04:09:43 crc kubenswrapper[4726]: I1004 04:09:43.504347 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:09:43 crc kubenswrapper[4726]: E1004 04:09:43.505976 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:09:49 crc kubenswrapper[4726]: I1004 04:09:49.057280 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4kwmz"] Oct 04 04:09:49 crc kubenswrapper[4726]: I1004 04:09:49.066902 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4kwmz"] Oct 04 04:09:50 crc kubenswrapper[4726]: I1004 04:09:50.512846 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0c7513-86d8-4b89-82ab-0a07c2d0d58a" path="/var/lib/kubelet/pods/fa0c7513-86d8-4b89-82ab-0a07c2d0d58a/volumes" Oct 04 04:09:51 crc kubenswrapper[4726]: I1004 04:09:51.038909 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-p7f5d"] Oct 04 04:09:51 crc kubenswrapper[4726]: I1004 04:09:51.060366 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-p7f5d"] Oct 04 04:09:52 crc kubenswrapper[4726]: I1004 04:09:52.516402 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480a81a8-69b1-4894-b36c-9f65a681af26" path="/var/lib/kubelet/pods/480a81a8-69b1-4894-b36c-9f65a681af26/volumes" Oct 04 04:09:56 crc kubenswrapper[4726]: I1004 04:09:56.504312 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:09:56 crc kubenswrapper[4726]: E1004 04:09:56.505795 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:07 crc kubenswrapper[4726]: I1004 04:10:07.503544 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:10:07 crc kubenswrapper[4726]: E1004 04:10:07.505328 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:18 crc kubenswrapper[4726]: I1004 04:10:18.503028 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:10:18 crc kubenswrapper[4726]: E1004 04:10:18.504237 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:20 crc kubenswrapper[4726]: I1004 04:10:20.849408 4726 generic.go:334] "Generic (PLEG): container finished" podID="4a4f5e93-bd73-42f9-8282-39e559ad08a7" containerID="ca77fc0b7cc20bf5f31370ecabb5883c71d492e92b1864dfb50a15217c3e288e" exitCode=0 Oct 04 04:10:20 crc kubenswrapper[4726]: I1004 04:10:20.849506 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" event={"ID":"4a4f5e93-bd73-42f9-8282-39e559ad08a7","Type":"ContainerDied","Data":"ca77fc0b7cc20bf5f31370ecabb5883c71d492e92b1864dfb50a15217c3e288e"} Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.395740 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.558361 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory\") pod \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.558490 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr2hr\" (UniqueName: \"kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr\") pod \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.558530 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key\") pod \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\" (UID: \"4a4f5e93-bd73-42f9-8282-39e559ad08a7\") " Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.565624 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr" (OuterVolumeSpecName: "kube-api-access-jr2hr") pod "4a4f5e93-bd73-42f9-8282-39e559ad08a7" (UID: "4a4f5e93-bd73-42f9-8282-39e559ad08a7"). InnerVolumeSpecName "kube-api-access-jr2hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.586425 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a4f5e93-bd73-42f9-8282-39e559ad08a7" (UID: "4a4f5e93-bd73-42f9-8282-39e559ad08a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.612089 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory" (OuterVolumeSpecName: "inventory") pod "4a4f5e93-bd73-42f9-8282-39e559ad08a7" (UID: "4a4f5e93-bd73-42f9-8282-39e559ad08a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.661635 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.661853 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr2hr\" (UniqueName: \"kubernetes.io/projected/4a4f5e93-bd73-42f9-8282-39e559ad08a7-kube-api-access-jr2hr\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.661883 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a4f5e93-bd73-42f9-8282-39e559ad08a7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.877847 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" event={"ID":"4a4f5e93-bd73-42f9-8282-39e559ad08a7","Type":"ContainerDied","Data":"cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df"} Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.877920 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbe214d781391c67820c8ddac5ec9a583ccefbbbafc416b010857024792f75df" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.878288 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.997332 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7"] Oct 04 04:10:22 crc kubenswrapper[4726]: E1004 04:10:22.997804 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a4f5e93-bd73-42f9-8282-39e559ad08a7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.997840 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a4f5e93-bd73-42f9-8282-39e559ad08a7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.998666 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a4f5e93-bd73-42f9-8282-39e559ad08a7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 04:10:22 crc kubenswrapper[4726]: I1004 04:10:22.999413 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.002156 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.002796 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.003604 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.005815 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.019672 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7"] Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.072688 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.072927 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcmr4\" (UniqueName: \"kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.073049 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.175563 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.175831 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcmr4\" (UniqueName: \"kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.175944 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.182706 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.185981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.217359 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcmr4\" (UniqueName: \"kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.320910 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:10:23 crc kubenswrapper[4726]: I1004 04:10:23.905853 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7"] Oct 04 04:10:24 crc kubenswrapper[4726]: I1004 04:10:24.903644 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" event={"ID":"ebf88812-a311-4c5a-8698-d50f551446f2","Type":"ContainerStarted","Data":"a1749e749c52f1cb681020604b50f4b678c75d7fb76bacc1dbad761ac90bf80c"} Oct 04 04:10:24 crc kubenswrapper[4726]: I1004 04:10:24.904022 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" event={"ID":"ebf88812-a311-4c5a-8698-d50f551446f2","Type":"ContainerStarted","Data":"7ad9bfeb0dfd470e211e597109b529a216766b1cdf8fe53043d7cf99d6b55a23"} Oct 04 04:10:24 crc kubenswrapper[4726]: I1004 04:10:24.946966 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" podStartSLOduration=2.51483679 podStartE2EDuration="2.946944102s" podCreationTimestamp="2025-10-04 04:10:22 +0000 UTC" firstStartedPulling="2025-10-04 04:10:23.910579619 +0000 UTC m=+1798.085202832" lastFinishedPulling="2025-10-04 04:10:24.342686891 +0000 UTC m=+1798.517310144" observedRunningTime="2025-10-04 04:10:24.93633689 +0000 UTC m=+1799.110960143" watchObservedRunningTime="2025-10-04 04:10:24.946944102 +0000 UTC m=+1799.121567325" Oct 04 04:10:29 crc kubenswrapper[4726]: I1004 04:10:29.503072 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:10:29 crc kubenswrapper[4726]: E1004 04:10:29.504486 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.710176 4726 scope.go:117] "RemoveContainer" containerID="cb09f557a83e3485f269200b4d32cf59e40c8f4027046324fa4e567da8894ba4" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.741448 4726 scope.go:117] "RemoveContainer" containerID="c97e6108ede611f87f19cbf31291f274b83a33880ce0078930832fa81e5600b3" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.826330 4726 scope.go:117] "RemoveContainer" containerID="5109ad53f6c94ca393fccc0dc0ee46d2f18a40f93b8cc828a299bef1970733c5" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.893987 4726 scope.go:117] "RemoveContainer" containerID="4d12216a7452549d7717e99df15b5b3f840797910fddcb7ad8a4f192b47eaf8c" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.923446 4726 scope.go:117] "RemoveContainer" containerID="06db8e023388e4bbb5cfc1bbf8ef2da4d68418cac61a93f0b40b5e26b32b7015" Oct 04 04:10:35 crc kubenswrapper[4726]: I1004 04:10:35.983515 4726 scope.go:117] "RemoveContainer" containerID="83100978dcd55367b1b844028edab28cf095b5d26392d709af8a1eb48d39876b" Oct 04 04:10:36 crc kubenswrapper[4726]: I1004 04:10:36.029608 4726 scope.go:117] "RemoveContainer" containerID="4420bcf7200cb7fd4a3cfea501e58e3e9b082cb9b59788d7278650033aa7230e" Oct 04 04:10:36 crc kubenswrapper[4726]: I1004 04:10:36.069050 4726 scope.go:117] "RemoveContainer" containerID="1b41125dcf3e5bb13869643424be34e7281b96914fe4b3dc5d14963d53ffce46" Oct 04 04:10:42 crc kubenswrapper[4726]: I1004 04:10:42.053567 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sn774"] Oct 04 04:10:42 crc kubenswrapper[4726]: I1004 04:10:42.061372 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sn774"] Oct 04 04:10:42 crc kubenswrapper[4726]: I1004 04:10:42.502734 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:10:42 crc kubenswrapper[4726]: E1004 04:10:42.503243 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:42 crc kubenswrapper[4726]: I1004 04:10:42.517986 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84f35fee-e3ff-49b7-980a-986aa0124565" path="/var/lib/kubelet/pods/84f35fee-e3ff-49b7-980a-986aa0124565/volumes" Oct 04 04:10:43 crc kubenswrapper[4726]: I1004 04:10:43.037787 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-q6284"] Oct 04 04:10:43 crc kubenswrapper[4726]: I1004 04:10:43.051500 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-q6284"] Oct 04 04:10:44 crc kubenswrapper[4726]: I1004 04:10:44.522236 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba7d46cf-02bd-4674-93e9-b339da1b568d" path="/var/lib/kubelet/pods/ba7d46cf-02bd-4674-93e9-b339da1b568d/volumes" Oct 04 04:10:49 crc kubenswrapper[4726]: I1004 04:10:49.074362 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xxqgf"] Oct 04 04:10:49 crc kubenswrapper[4726]: I1004 04:10:49.089335 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-j9jwl"] Oct 04 04:10:49 crc kubenswrapper[4726]: I1004 04:10:49.102983 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xxqgf"] Oct 04 04:10:49 crc kubenswrapper[4726]: I1004 04:10:49.111669 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-j9jwl"] Oct 04 04:10:50 crc kubenswrapper[4726]: I1004 04:10:50.522257 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90894cee-ffab-4485-a221-703de154182f" path="/var/lib/kubelet/pods/90894cee-ffab-4485-a221-703de154182f/volumes" Oct 04 04:10:50 crc kubenswrapper[4726]: I1004 04:10:50.524054 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf9ee75-66f7-4371-9adc-c0ad0249698c" path="/var/lib/kubelet/pods/caf9ee75-66f7-4371-9adc-c0ad0249698c/volumes" Oct 04 04:10:55 crc kubenswrapper[4726]: I1004 04:10:55.503439 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:10:55 crc kubenswrapper[4726]: E1004 04:10:55.504290 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:10:59 crc kubenswrapper[4726]: I1004 04:10:59.043663 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-72jth"] Oct 04 04:10:59 crc kubenswrapper[4726]: I1004 04:10:59.060232 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-72jth"] Oct 04 04:11:00 crc kubenswrapper[4726]: I1004 04:11:00.513149 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f5261e9-4ede-4c36-83cd-58cfe6176333" path="/var/lib/kubelet/pods/1f5261e9-4ede-4c36-83cd-58cfe6176333/volumes" Oct 04 04:11:07 crc kubenswrapper[4726]: I1004 04:11:07.502287 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:11:07 crc kubenswrapper[4726]: E1004 04:11:07.503139 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:11:22 crc kubenswrapper[4726]: I1004 04:11:22.502926 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:11:22 crc kubenswrapper[4726]: E1004 04:11:22.504301 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:11:35 crc kubenswrapper[4726]: I1004 04:11:35.502499 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:11:35 crc kubenswrapper[4726]: I1004 04:11:35.718767 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb"} Oct 04 04:11:36 crc kubenswrapper[4726]: I1004 04:11:36.279386 4726 scope.go:117] "RemoveContainer" containerID="582be0e212516c976b60d9788c7eb4b00289ff683ab750e1b4481871333cf4a9" Oct 04 04:11:36 crc kubenswrapper[4726]: I1004 04:11:36.344661 4726 scope.go:117] "RemoveContainer" containerID="36ed94a606fbe6d95569c338bfc829a027d7fbb9bdf4294663ce49b3b96b58b0" Oct 04 04:11:36 crc kubenswrapper[4726]: I1004 04:11:36.419799 4726 scope.go:117] "RemoveContainer" containerID="e5c4b24a358c073895352df74618c137ab92ae661ae669c963644d144ab8cd7f" Oct 04 04:11:36 crc kubenswrapper[4726]: I1004 04:11:36.477968 4726 scope.go:117] "RemoveContainer" containerID="dcfe4393bce42582a93d2a1065b96d65316474d253fdb4783fb43db8f0509c39" Oct 04 04:11:36 crc kubenswrapper[4726]: I1004 04:11:36.533499 4726 scope.go:117] "RemoveContainer" containerID="13a6fc74ee64bc5598209bdc886cc65367098b521dea8cd60ae798ba13febe78" Oct 04 04:11:41 crc kubenswrapper[4726]: I1004 04:11:41.790571 4726 generic.go:334] "Generic (PLEG): container finished" podID="ebf88812-a311-4c5a-8698-d50f551446f2" containerID="a1749e749c52f1cb681020604b50f4b678c75d7fb76bacc1dbad761ac90bf80c" exitCode=0 Oct 04 04:11:41 crc kubenswrapper[4726]: I1004 04:11:41.790783 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" event={"ID":"ebf88812-a311-4c5a-8698-d50f551446f2","Type":"ContainerDied","Data":"a1749e749c52f1cb681020604b50f4b678c75d7fb76bacc1dbad761ac90bf80c"} Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.062208 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-x8ln8"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.071508 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jdcns"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.079646 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8skgc"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.089175 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-x8ln8"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.097995 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8skgc"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.111060 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jdcns"] Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.523315 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ce587ac-9eb6-427d-b15c-399d40d440b9" path="/var/lib/kubelet/pods/6ce587ac-9eb6-427d-b15c-399d40d440b9/volumes" Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.524637 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9830424-4983-4606-bf56-d6e0819d9eec" path="/var/lib/kubelet/pods/a9830424-4983-4606-bf56-d6e0819d9eec/volumes" Oct 04 04:11:42 crc kubenswrapper[4726]: I1004 04:11:42.525884 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70f6012-daec-42cb-b962-2864279cf632" path="/var/lib/kubelet/pods/e70f6012-daec-42cb-b962-2864279cf632/volumes" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.257961 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.349306 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory\") pod \"ebf88812-a311-4c5a-8698-d50f551446f2\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.350491 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcmr4\" (UniqueName: \"kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4\") pod \"ebf88812-a311-4c5a-8698-d50f551446f2\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.351992 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key\") pod \"ebf88812-a311-4c5a-8698-d50f551446f2\" (UID: \"ebf88812-a311-4c5a-8698-d50f551446f2\") " Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.358212 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4" (OuterVolumeSpecName: "kube-api-access-qcmr4") pod "ebf88812-a311-4c5a-8698-d50f551446f2" (UID: "ebf88812-a311-4c5a-8698-d50f551446f2"). InnerVolumeSpecName "kube-api-access-qcmr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.384000 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory" (OuterVolumeSpecName: "inventory") pod "ebf88812-a311-4c5a-8698-d50f551446f2" (UID: "ebf88812-a311-4c5a-8698-d50f551446f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.385464 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ebf88812-a311-4c5a-8698-d50f551446f2" (UID: "ebf88812-a311-4c5a-8698-d50f551446f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.454409 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.454684 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ebf88812-a311-4c5a-8698-d50f551446f2-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.454765 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcmr4\" (UniqueName: \"kubernetes.io/projected/ebf88812-a311-4c5a-8698-d50f551446f2-kube-api-access-qcmr4\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.819075 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" event={"ID":"ebf88812-a311-4c5a-8698-d50f551446f2","Type":"ContainerDied","Data":"7ad9bfeb0dfd470e211e597109b529a216766b1cdf8fe53043d7cf99d6b55a23"} Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.819508 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ad9bfeb0dfd470e211e597109b529a216766b1cdf8fe53043d7cf99d6b55a23" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.819209 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.924539 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w"] Oct 04 04:11:43 crc kubenswrapper[4726]: E1004 04:11:43.924993 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf88812-a311-4c5a-8698-d50f551446f2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.925013 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf88812-a311-4c5a-8698-d50f551446f2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.925289 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf88812-a311-4c5a-8698-d50f551446f2" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.926058 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.928515 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.931150 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.931765 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.937522 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.939125 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w"] Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.967617 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.967828 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:43 crc kubenswrapper[4726]: I1004 04:11:43.967887 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd4jf\" (UniqueName: \"kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.069690 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.069737 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd4jf\" (UniqueName: \"kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.069841 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.074016 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.074134 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.107314 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd4jf\" (UniqueName: \"kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.247830 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:44 crc kubenswrapper[4726]: I1004 04:11:44.847770 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w"] Oct 04 04:11:44 crc kubenswrapper[4726]: W1004 04:11:44.850297 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaffede2f_dbd9_44c4_ae15_af1fb96ce9f6.slice/crio-2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128 WatchSource:0}: Error finding container 2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128: Status 404 returned error can't find the container with id 2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128 Oct 04 04:11:45 crc kubenswrapper[4726]: I1004 04:11:45.840209 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" event={"ID":"affede2f-dbd9-44c4-ae15-af1fb96ce9f6","Type":"ContainerStarted","Data":"86a9ccf9d631f829804b768b1e65710d8a2cc94abe9d657712871128863f7c41"} Oct 04 04:11:45 crc kubenswrapper[4726]: I1004 04:11:45.840995 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" event={"ID":"affede2f-dbd9-44c4-ae15-af1fb96ce9f6","Type":"ContainerStarted","Data":"2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128"} Oct 04 04:11:45 crc kubenswrapper[4726]: I1004 04:11:45.868780 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" podStartSLOduration=2.426636382 podStartE2EDuration="2.86875727s" podCreationTimestamp="2025-10-04 04:11:43 +0000 UTC" firstStartedPulling="2025-10-04 04:11:44.853682177 +0000 UTC m=+1879.028305420" lastFinishedPulling="2025-10-04 04:11:45.295803065 +0000 UTC m=+1879.470426308" observedRunningTime="2025-10-04 04:11:45.866462791 +0000 UTC m=+1880.041086034" watchObservedRunningTime="2025-10-04 04:11:45.86875727 +0000 UTC m=+1880.043380503" Oct 04 04:11:50 crc kubenswrapper[4726]: I1004 04:11:50.894388 4726 generic.go:334] "Generic (PLEG): container finished" podID="affede2f-dbd9-44c4-ae15-af1fb96ce9f6" containerID="86a9ccf9d631f829804b768b1e65710d8a2cc94abe9d657712871128863f7c41" exitCode=0 Oct 04 04:11:50 crc kubenswrapper[4726]: I1004 04:11:50.894492 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" event={"ID":"affede2f-dbd9-44c4-ae15-af1fb96ce9f6","Type":"ContainerDied","Data":"86a9ccf9d631f829804b768b1e65710d8a2cc94abe9d657712871128863f7c41"} Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.057017 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-d3fb-account-create-7v5gc"] Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.083115 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3795-account-create-b95q8"] Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.099127 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3795-account-create-b95q8"] Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.110149 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-d3fb-account-create-7v5gc"] Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.394587 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.512814 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="200b106d-3b99-40c1-9368-32370f56d01a" path="/var/lib/kubelet/pods/200b106d-3b99-40c1-9368-32370f56d01a/volumes" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.513334 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2a19f3-d51c-4ea0-a16c-3b0bad069a69" path="/var/lib/kubelet/pods/7e2a19f3-d51c-4ea0-a16c-3b0bad069a69/volumes" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.569533 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory\") pod \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.569723 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd4jf\" (UniqueName: \"kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf\") pod \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.569954 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key\") pod \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\" (UID: \"affede2f-dbd9-44c4-ae15-af1fb96ce9f6\") " Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.576724 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf" (OuterVolumeSpecName: "kube-api-access-zd4jf") pod "affede2f-dbd9-44c4-ae15-af1fb96ce9f6" (UID: "affede2f-dbd9-44c4-ae15-af1fb96ce9f6"). InnerVolumeSpecName "kube-api-access-zd4jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.600372 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "affede2f-dbd9-44c4-ae15-af1fb96ce9f6" (UID: "affede2f-dbd9-44c4-ae15-af1fb96ce9f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.616409 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory" (OuterVolumeSpecName: "inventory") pod "affede2f-dbd9-44c4-ae15-af1fb96ce9f6" (UID: "affede2f-dbd9-44c4-ae15-af1fb96ce9f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.672915 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.672951 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd4jf\" (UniqueName: \"kubernetes.io/projected/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-kube-api-access-zd4jf\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.672963 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/affede2f-dbd9-44c4-ae15-af1fb96ce9f6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.917536 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" event={"ID":"affede2f-dbd9-44c4-ae15-af1fb96ce9f6","Type":"ContainerDied","Data":"2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128"} Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.917584 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ba30d746872097549b0bfae0f11c8d0ee9d5bc73ccf3c4370e1431007188128" Oct 04 04:11:52 crc kubenswrapper[4726]: I1004 04:11:52.917611 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.018230 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc"] Oct 04 04:11:53 crc kubenswrapper[4726]: E1004 04:11:53.018938 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affede2f-dbd9-44c4-ae15-af1fb96ce9f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.018987 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="affede2f-dbd9-44c4-ae15-af1fb96ce9f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.019527 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="affede2f-dbd9-44c4-ae15-af1fb96ce9f6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.020672 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.024027 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.024081 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.024142 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.024399 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.028712 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc"] Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.071068 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-54bd-account-create-698qk"] Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.080226 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvqss\" (UniqueName: \"kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.080573 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.080713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.082793 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-54bd-account-create-698qk"] Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.182493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvqss\" (UniqueName: \"kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.182760 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.182885 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.187747 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.189766 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.202261 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvqss\" (UniqueName: \"kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-v85jc\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.342216 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.672258 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc"] Oct 04 04:11:53 crc kubenswrapper[4726]: W1004 04:11:53.675186 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1faabfae_0c50_425d_af5b_529a69ee6791.slice/crio-318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c WatchSource:0}: Error finding container 318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c: Status 404 returned error can't find the container with id 318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c Oct 04 04:11:53 crc kubenswrapper[4726]: I1004 04:11:53.929660 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" event={"ID":"1faabfae-0c50-425d-af5b-529a69ee6791","Type":"ContainerStarted","Data":"318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c"} Oct 04 04:11:54 crc kubenswrapper[4726]: I1004 04:11:54.517309 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b65ae8b8-8121-45de-9647-f90d71a923c9" path="/var/lib/kubelet/pods/b65ae8b8-8121-45de-9647-f90d71a923c9/volumes" Oct 04 04:11:54 crc kubenswrapper[4726]: I1004 04:11:54.950005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" event={"ID":"1faabfae-0c50-425d-af5b-529a69ee6791","Type":"ContainerStarted","Data":"2e367c89d52262891198e0f2a7261c6b3d93cd4dd8d2519280c09d5fde6a642c"} Oct 04 04:11:54 crc kubenswrapper[4726]: I1004 04:11:54.969988 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" podStartSLOduration=2.565070859 podStartE2EDuration="2.969968093s" podCreationTimestamp="2025-10-04 04:11:52 +0000 UTC" firstStartedPulling="2025-10-04 04:11:53.67800523 +0000 UTC m=+1887.852628453" lastFinishedPulling="2025-10-04 04:11:54.082902434 +0000 UTC m=+1888.257525687" observedRunningTime="2025-10-04 04:11:54.9652137 +0000 UTC m=+1889.139836943" watchObservedRunningTime="2025-10-04 04:11:54.969968093 +0000 UTC m=+1889.144591306" Oct 04 04:12:15 crc kubenswrapper[4726]: I1004 04:12:15.047485 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7t6dt"] Oct 04 04:12:15 crc kubenswrapper[4726]: I1004 04:12:15.063259 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7t6dt"] Oct 04 04:12:16 crc kubenswrapper[4726]: I1004 04:12:16.520225 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88100d2e-2e91-49d5-8f3f-f51864942282" path="/var/lib/kubelet/pods/88100d2e-2e91-49d5-8f3f-f51864942282/volumes" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.693973 4726 scope.go:117] "RemoveContainer" containerID="c71c59d33610947f261bf22c3e0bd3659db271acc042bdef71c860d35774e971" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.729477 4726 scope.go:117] "RemoveContainer" containerID="a2f202197b419f3f771ab7df93d402e98b1f68b10496add4d5d2e3d28bb8e634" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.799426 4726 scope.go:117] "RemoveContainer" containerID="8acbc066cb63c6b186f8b40efaade29ed567885912707f43633abe27f145e77d" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.889365 4726 scope.go:117] "RemoveContainer" containerID="dec3a6a5787a134fadb9a0c8f2198464c5bd4617a6bf335d16b9dcb3c027af1b" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.915313 4726 scope.go:117] "RemoveContainer" containerID="2d49ea6ae564769a8f4f839243651683853779020afb3cb6050802db1fbd5c5c" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.954871 4726 scope.go:117] "RemoveContainer" containerID="f1b3c30f91cca95b2b19725869571981aecb3fb57d2281b1f9d07e7414ea2ed7" Oct 04 04:12:36 crc kubenswrapper[4726]: I1004 04:12:36.999143 4726 scope.go:117] "RemoveContainer" containerID="8052916d2328c64bfe226e9102cc5b3b18c541bec7577f8fbe032b05cbc26c31" Oct 04 04:12:37 crc kubenswrapper[4726]: I1004 04:12:37.060536 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-m8rwl"] Oct 04 04:12:37 crc kubenswrapper[4726]: I1004 04:12:37.071956 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-m8rwl"] Oct 04 04:12:38 crc kubenswrapper[4726]: I1004 04:12:38.528599 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3dba01d-46b8-407b-9622-5fd1534b9edc" path="/var/lib/kubelet/pods/b3dba01d-46b8-407b-9622-5fd1534b9edc/volumes" Oct 04 04:12:39 crc kubenswrapper[4726]: I1004 04:12:39.514169 4726 generic.go:334] "Generic (PLEG): container finished" podID="1faabfae-0c50-425d-af5b-529a69ee6791" containerID="2e367c89d52262891198e0f2a7261c6b3d93cd4dd8d2519280c09d5fde6a642c" exitCode=0 Oct 04 04:12:39 crc kubenswrapper[4726]: I1004 04:12:39.514225 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" event={"ID":"1faabfae-0c50-425d-af5b-529a69ee6791","Type":"ContainerDied","Data":"2e367c89d52262891198e0f2a7261c6b3d93cd4dd8d2519280c09d5fde6a642c"} Oct 04 04:12:40 crc kubenswrapper[4726]: I1004 04:12:40.043157 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x69lg"] Oct 04 04:12:40 crc kubenswrapper[4726]: I1004 04:12:40.058871 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x69lg"] Oct 04 04:12:40 crc kubenswrapper[4726]: I1004 04:12:40.521608 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e8ff110-c94f-4dc0-8420-e00cff505a20" path="/var/lib/kubelet/pods/0e8ff110-c94f-4dc0-8420-e00cff505a20/volumes" Oct 04 04:12:40 crc kubenswrapper[4726]: I1004 04:12:40.983969 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.125845 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory\") pod \"1faabfae-0c50-425d-af5b-529a69ee6791\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.127591 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key\") pod \"1faabfae-0c50-425d-af5b-529a69ee6791\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.127787 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvqss\" (UniqueName: \"kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss\") pod \"1faabfae-0c50-425d-af5b-529a69ee6791\" (UID: \"1faabfae-0c50-425d-af5b-529a69ee6791\") " Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.133823 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss" (OuterVolumeSpecName: "kube-api-access-zvqss") pod "1faabfae-0c50-425d-af5b-529a69ee6791" (UID: "1faabfae-0c50-425d-af5b-529a69ee6791"). InnerVolumeSpecName "kube-api-access-zvqss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.153284 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory" (OuterVolumeSpecName: "inventory") pod "1faabfae-0c50-425d-af5b-529a69ee6791" (UID: "1faabfae-0c50-425d-af5b-529a69ee6791"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.159758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1faabfae-0c50-425d-af5b-529a69ee6791" (UID: "1faabfae-0c50-425d-af5b-529a69ee6791"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.230049 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.230268 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvqss\" (UniqueName: \"kubernetes.io/projected/1faabfae-0c50-425d-af5b-529a69ee6791-kube-api-access-zvqss\") on node \"crc\" DevicePath \"\"" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.230365 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1faabfae-0c50-425d-af5b-529a69ee6791-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.542193 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" event={"ID":"1faabfae-0c50-425d-af5b-529a69ee6791","Type":"ContainerDied","Data":"318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c"} Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.542270 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="318f05ba66b6af68b2dd1c1182094ac985fdbe8b3f0e885d51afd207edd6035c" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.542330 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-v85jc" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.672198 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q"] Oct 04 04:12:41 crc kubenswrapper[4726]: E1004 04:12:41.672813 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1faabfae-0c50-425d-af5b-529a69ee6791" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.672837 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="1faabfae-0c50-425d-af5b-529a69ee6791" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.673213 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="1faabfae-0c50-425d-af5b-529a69ee6791" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.674502 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.677967 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.678588 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.680316 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.683490 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.686279 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q"] Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.841713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.841810 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6f4g\" (UniqueName: \"kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.842363 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.945999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.946085 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.946128 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6f4g\" (UniqueName: \"kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.952640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.953317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.977507 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6f4g\" (UniqueName: \"kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:41 crc kubenswrapper[4726]: I1004 04:12:41.998550 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:12:42 crc kubenswrapper[4726]: I1004 04:12:42.433840 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q"] Oct 04 04:12:42 crc kubenswrapper[4726]: I1004 04:12:42.553618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" event={"ID":"13859a35-3c01-4314-a416-0d3d04828cdc","Type":"ContainerStarted","Data":"68eecf5bf7d60b656931fb7f20c0677e1023774ff150955084c4ac3b2392ce57"} Oct 04 04:12:43 crc kubenswrapper[4726]: I1004 04:12:43.568313 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" event={"ID":"13859a35-3c01-4314-a416-0d3d04828cdc","Type":"ContainerStarted","Data":"fc390594631c4560423fd5edb74f6b25b647b52b0b6d48fa96dfedcfd20f7b6e"} Oct 04 04:12:43 crc kubenswrapper[4726]: I1004 04:12:43.604492 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" podStartSLOduration=2.15904407 podStartE2EDuration="2.604461593s" podCreationTimestamp="2025-10-04 04:12:41 +0000 UTC" firstStartedPulling="2025-10-04 04:12:42.436932823 +0000 UTC m=+1936.611556056" lastFinishedPulling="2025-10-04 04:12:42.882350326 +0000 UTC m=+1937.056973579" observedRunningTime="2025-10-04 04:12:43.589505806 +0000 UTC m=+1937.764129059" watchObservedRunningTime="2025-10-04 04:12:43.604461593 +0000 UTC m=+1937.779084846" Oct 04 04:13:22 crc kubenswrapper[4726]: I1004 04:13:22.072962 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gbbrr"] Oct 04 04:13:22 crc kubenswrapper[4726]: I1004 04:13:22.084854 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gbbrr"] Oct 04 04:13:22 crc kubenswrapper[4726]: I1004 04:13:22.520453 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd02f9ea-7639-46f4-b330-8a5cb3864e9d" path="/var/lib/kubelet/pods/cd02f9ea-7639-46f4-b330-8a5cb3864e9d/volumes" Oct 04 04:13:37 crc kubenswrapper[4726]: I1004 04:13:37.229486 4726 scope.go:117] "RemoveContainer" containerID="624afb6db3583fb9ac26f7e2920aa28fced0811676a72d315ff1c88c5ac52b21" Oct 04 04:13:37 crc kubenswrapper[4726]: I1004 04:13:37.295097 4726 scope.go:117] "RemoveContainer" containerID="491f5e8a8a85b7bc0611134218b495881366a5a89117ef627d978935edc8100e" Oct 04 04:13:37 crc kubenswrapper[4726]: I1004 04:13:37.360098 4726 scope.go:117] "RemoveContainer" containerID="104e20c2aa36829f1dab47759e7207c41c04cde7a1b78606cd3d7ef8105cc84a" Oct 04 04:13:42 crc kubenswrapper[4726]: I1004 04:13:42.281489 4726 generic.go:334] "Generic (PLEG): container finished" podID="13859a35-3c01-4314-a416-0d3d04828cdc" containerID="fc390594631c4560423fd5edb74f6b25b647b52b0b6d48fa96dfedcfd20f7b6e" exitCode=2 Oct 04 04:13:42 crc kubenswrapper[4726]: I1004 04:13:42.281635 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" event={"ID":"13859a35-3c01-4314-a416-0d3d04828cdc","Type":"ContainerDied","Data":"fc390594631c4560423fd5edb74f6b25b647b52b0b6d48fa96dfedcfd20f7b6e"} Oct 04 04:13:43 crc kubenswrapper[4726]: I1004 04:13:43.914765 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.064406 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory\") pod \"13859a35-3c01-4314-a416-0d3d04828cdc\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.064532 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6f4g\" (UniqueName: \"kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g\") pod \"13859a35-3c01-4314-a416-0d3d04828cdc\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.064789 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key\") pod \"13859a35-3c01-4314-a416-0d3d04828cdc\" (UID: \"13859a35-3c01-4314-a416-0d3d04828cdc\") " Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.073503 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g" (OuterVolumeSpecName: "kube-api-access-q6f4g") pod "13859a35-3c01-4314-a416-0d3d04828cdc" (UID: "13859a35-3c01-4314-a416-0d3d04828cdc"). InnerVolumeSpecName "kube-api-access-q6f4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.116282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "13859a35-3c01-4314-a416-0d3d04828cdc" (UID: "13859a35-3c01-4314-a416-0d3d04828cdc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.117996 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory" (OuterVolumeSpecName: "inventory") pod "13859a35-3c01-4314-a416-0d3d04828cdc" (UID: "13859a35-3c01-4314-a416-0d3d04828cdc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.167692 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.167736 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6f4g\" (UniqueName: \"kubernetes.io/projected/13859a35-3c01-4314-a416-0d3d04828cdc-kube-api-access-q6f4g\") on node \"crc\" DevicePath \"\"" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.167754 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13859a35-3c01-4314-a416-0d3d04828cdc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.306969 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" event={"ID":"13859a35-3c01-4314-a416-0d3d04828cdc","Type":"ContainerDied","Data":"68eecf5bf7d60b656931fb7f20c0677e1023774ff150955084c4ac3b2392ce57"} Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.307341 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68eecf5bf7d60b656931fb7f20c0677e1023774ff150955084c4ac3b2392ce57" Oct 04 04:13:44 crc kubenswrapper[4726]: I1004 04:13:44.307054 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.023764 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5"] Oct 04 04:13:51 crc kubenswrapper[4726]: E1004 04:13:51.024670 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13859a35-3c01-4314-a416-0d3d04828cdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.024683 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="13859a35-3c01-4314-a416-0d3d04828cdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.024837 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="13859a35-3c01-4314-a416-0d3d04828cdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.025446 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.027928 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.029456 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.032816 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.036088 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.072430 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5"] Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.128926 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.129046 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvzgk\" (UniqueName: \"kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.129116 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.231570 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.231663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvzgk\" (UniqueName: \"kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.231785 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.238223 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.239527 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.251080 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvzgk\" (UniqueName: \"kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.347156 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.946541 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5"] Oct 04 04:13:51 crc kubenswrapper[4726]: I1004 04:13:51.972369 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:13:52 crc kubenswrapper[4726]: I1004 04:13:52.388560 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" event={"ID":"dec26676-436d-4d9a-8337-0c9ac86c6168","Type":"ContainerStarted","Data":"41aaa73f24e0c610074573943fc86a89ccd2344d3ba0c9053b3647d7e9cb99c6"} Oct 04 04:13:53 crc kubenswrapper[4726]: I1004 04:13:53.401281 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" event={"ID":"dec26676-436d-4d9a-8337-0c9ac86c6168","Type":"ContainerStarted","Data":"3f6470ea0650ab7e60dc0474cb1fb99109868ad8213ab8cb9a7b2c6793945b7b"} Oct 04 04:13:53 crc kubenswrapper[4726]: I1004 04:13:53.423189 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" podStartSLOduration=1.880502414 podStartE2EDuration="2.423161565s" podCreationTimestamp="2025-10-04 04:13:51 +0000 UTC" firstStartedPulling="2025-10-04 04:13:51.971967935 +0000 UTC m=+2006.146591178" lastFinishedPulling="2025-10-04 04:13:52.514627116 +0000 UTC m=+2006.689250329" observedRunningTime="2025-10-04 04:13:53.422063986 +0000 UTC m=+2007.596687239" watchObservedRunningTime="2025-10-04 04:13:53.423161565 +0000 UTC m=+2007.597784808" Oct 04 04:14:04 crc kubenswrapper[4726]: I1004 04:14:04.187968 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:14:04 crc kubenswrapper[4726]: I1004 04:14:04.188769 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.099164 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.102516 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.125583 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.210825 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.210954 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmb66\" (UniqueName: \"kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.211070 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.313420 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.313716 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmb66\" (UniqueName: \"kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.314282 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.314574 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.315043 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.337253 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmb66\" (UniqueName: \"kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66\") pod \"community-operators-dfnqr\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:23 crc kubenswrapper[4726]: I1004 04:14:23.455770 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:24 crc kubenswrapper[4726]: I1004 04:14:24.016445 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:24 crc kubenswrapper[4726]: I1004 04:14:24.744838 4726 generic.go:334] "Generic (PLEG): container finished" podID="a48f3207-0197-4361-958b-b7c491b3495c" containerID="cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5" exitCode=0 Oct 04 04:14:24 crc kubenswrapper[4726]: I1004 04:14:24.744904 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerDied","Data":"cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5"} Oct 04 04:14:24 crc kubenswrapper[4726]: I1004 04:14:24.744946 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerStarted","Data":"80adab00f102d0d6fe5799cf19d859021165748ee804136096baf21b39ecc86b"} Oct 04 04:14:25 crc kubenswrapper[4726]: I1004 04:14:25.756861 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerStarted","Data":"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537"} Oct 04 04:14:26 crc kubenswrapper[4726]: I1004 04:14:26.771349 4726 generic.go:334] "Generic (PLEG): container finished" podID="a48f3207-0197-4361-958b-b7c491b3495c" containerID="3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537" exitCode=0 Oct 04 04:14:26 crc kubenswrapper[4726]: I1004 04:14:26.771417 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerDied","Data":"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537"} Oct 04 04:14:27 crc kubenswrapper[4726]: I1004 04:14:27.780822 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerStarted","Data":"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6"} Oct 04 04:14:27 crc kubenswrapper[4726]: I1004 04:14:27.798706 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dfnqr" podStartSLOduration=2.235711375 podStartE2EDuration="4.798687304s" podCreationTimestamp="2025-10-04 04:14:23 +0000 UTC" firstStartedPulling="2025-10-04 04:14:24.747975539 +0000 UTC m=+2038.922598792" lastFinishedPulling="2025-10-04 04:14:27.310951498 +0000 UTC m=+2041.485574721" observedRunningTime="2025-10-04 04:14:27.796150138 +0000 UTC m=+2041.970773361" watchObservedRunningTime="2025-10-04 04:14:27.798687304 +0000 UTC m=+2041.973310517" Oct 04 04:14:33 crc kubenswrapper[4726]: I1004 04:14:33.456684 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:33 crc kubenswrapper[4726]: I1004 04:14:33.457457 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:33 crc kubenswrapper[4726]: I1004 04:14:33.548137 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:33 crc kubenswrapper[4726]: I1004 04:14:33.936005 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:34 crc kubenswrapper[4726]: I1004 04:14:34.003504 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:34 crc kubenswrapper[4726]: I1004 04:14:34.189013 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:14:34 crc kubenswrapper[4726]: I1004 04:14:34.189150 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:14:35 crc kubenswrapper[4726]: I1004 04:14:35.881418 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dfnqr" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="registry-server" containerID="cri-o://0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6" gracePeriod=2 Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.463927 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.529355 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities\") pod \"a48f3207-0197-4361-958b-b7c491b3495c\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.529446 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content\") pod \"a48f3207-0197-4361-958b-b7c491b3495c\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.529537 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmb66\" (UniqueName: \"kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66\") pod \"a48f3207-0197-4361-958b-b7c491b3495c\" (UID: \"a48f3207-0197-4361-958b-b7c491b3495c\") " Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.530232 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities" (OuterVolumeSpecName: "utilities") pod "a48f3207-0197-4361-958b-b7c491b3495c" (UID: "a48f3207-0197-4361-958b-b7c491b3495c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.530893 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.538805 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66" (OuterVolumeSpecName: "kube-api-access-jmb66") pod "a48f3207-0197-4361-958b-b7c491b3495c" (UID: "a48f3207-0197-4361-958b-b7c491b3495c"). InnerVolumeSpecName "kube-api-access-jmb66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.578741 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a48f3207-0197-4361-958b-b7c491b3495c" (UID: "a48f3207-0197-4361-958b-b7c491b3495c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.634060 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a48f3207-0197-4361-958b-b7c491b3495c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.634410 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmb66\" (UniqueName: \"kubernetes.io/projected/a48f3207-0197-4361-958b-b7c491b3495c-kube-api-access-jmb66\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.895557 4726 generic.go:334] "Generic (PLEG): container finished" podID="a48f3207-0197-4361-958b-b7c491b3495c" containerID="0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6" exitCode=0 Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.895617 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerDied","Data":"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6"} Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.895658 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dfnqr" event={"ID":"a48f3207-0197-4361-958b-b7c491b3495c","Type":"ContainerDied","Data":"80adab00f102d0d6fe5799cf19d859021165748ee804136096baf21b39ecc86b"} Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.895690 4726 scope.go:117] "RemoveContainer" containerID="0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.895723 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dfnqr" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.932771 4726 scope.go:117] "RemoveContainer" containerID="3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537" Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.961195 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.968702 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dfnqr"] Oct 04 04:14:36 crc kubenswrapper[4726]: I1004 04:14:36.973610 4726 scope.go:117] "RemoveContainer" containerID="cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.019635 4726 scope.go:117] "RemoveContainer" containerID="0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6" Oct 04 04:14:37 crc kubenswrapper[4726]: E1004 04:14:37.019967 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6\": container with ID starting with 0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6 not found: ID does not exist" containerID="0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.020000 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6"} err="failed to get container status \"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6\": rpc error: code = NotFound desc = could not find container \"0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6\": container with ID starting with 0c56096f042207188de77be2043098ad32790e81e86adf85a65684974cd75ad6 not found: ID does not exist" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.020020 4726 scope.go:117] "RemoveContainer" containerID="3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537" Oct 04 04:14:37 crc kubenswrapper[4726]: E1004 04:14:37.020277 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537\": container with ID starting with 3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537 not found: ID does not exist" containerID="3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.020311 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537"} err="failed to get container status \"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537\": rpc error: code = NotFound desc = could not find container \"3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537\": container with ID starting with 3f6655db76b019fc8faebcdc71b9ab7ce90a1a337a81b361ac43d5873938b537 not found: ID does not exist" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.020324 4726 scope.go:117] "RemoveContainer" containerID="cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5" Oct 04 04:14:37 crc kubenswrapper[4726]: E1004 04:14:37.020582 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5\": container with ID starting with cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5 not found: ID does not exist" containerID="cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5" Oct 04 04:14:37 crc kubenswrapper[4726]: I1004 04:14:37.020610 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5"} err="failed to get container status \"cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5\": rpc error: code = NotFound desc = could not find container \"cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5\": container with ID starting with cf30d269fc521302b41116cab88841abcc503e0ad66c54fa738b2e33c59f9af5 not found: ID does not exist" Oct 04 04:14:38 crc kubenswrapper[4726]: I1004 04:14:38.518138 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a48f3207-0197-4361-958b-b7c491b3495c" path="/var/lib/kubelet/pods/a48f3207-0197-4361-958b-b7c491b3495c/volumes" Oct 04 04:14:50 crc kubenswrapper[4726]: I1004 04:14:50.075749 4726 generic.go:334] "Generic (PLEG): container finished" podID="dec26676-436d-4d9a-8337-0c9ac86c6168" containerID="3f6470ea0650ab7e60dc0474cb1fb99109868ad8213ab8cb9a7b2c6793945b7b" exitCode=0 Oct 04 04:14:50 crc kubenswrapper[4726]: I1004 04:14:50.075890 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" event={"ID":"dec26676-436d-4d9a-8337-0c9ac86c6168","Type":"ContainerDied","Data":"3f6470ea0650ab7e60dc0474cb1fb99109868ad8213ab8cb9a7b2c6793945b7b"} Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.586507 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.681569 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvzgk\" (UniqueName: \"kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk\") pod \"dec26676-436d-4d9a-8337-0c9ac86c6168\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.681661 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key\") pod \"dec26676-436d-4d9a-8337-0c9ac86c6168\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.681826 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory\") pod \"dec26676-436d-4d9a-8337-0c9ac86c6168\" (UID: \"dec26676-436d-4d9a-8337-0c9ac86c6168\") " Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.688375 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk" (OuterVolumeSpecName: "kube-api-access-tvzgk") pod "dec26676-436d-4d9a-8337-0c9ac86c6168" (UID: "dec26676-436d-4d9a-8337-0c9ac86c6168"). InnerVolumeSpecName "kube-api-access-tvzgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.709906 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory" (OuterVolumeSpecName: "inventory") pod "dec26676-436d-4d9a-8337-0c9ac86c6168" (UID: "dec26676-436d-4d9a-8337-0c9ac86c6168"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.730740 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dec26676-436d-4d9a-8337-0c9ac86c6168" (UID: "dec26676-436d-4d9a-8337-0c9ac86c6168"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.784016 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.784058 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvzgk\" (UniqueName: \"kubernetes.io/projected/dec26676-436d-4d9a-8337-0c9ac86c6168-kube-api-access-tvzgk\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:51 crc kubenswrapper[4726]: I1004 04:14:51.784073 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dec26676-436d-4d9a-8337-0c9ac86c6168-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.115511 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" event={"ID":"dec26676-436d-4d9a-8337-0c9ac86c6168","Type":"ContainerDied","Data":"41aaa73f24e0c610074573943fc86a89ccd2344d3ba0c9053b3647d7e9cb99c6"} Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.115556 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41aaa73f24e0c610074573943fc86a89ccd2344d3ba0c9053b3647d7e9cb99c6" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.115634 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.252751 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mjmvr"] Oct 04 04:14:52 crc kubenswrapper[4726]: E1004 04:14:52.253201 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dec26676-436d-4d9a-8337-0c9ac86c6168" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253220 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="dec26676-436d-4d9a-8337-0c9ac86c6168" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:14:52 crc kubenswrapper[4726]: E1004 04:14:52.253241 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="extract-utilities" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253249 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="extract-utilities" Oct 04 04:14:52 crc kubenswrapper[4726]: E1004 04:14:52.253264 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="extract-content" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253270 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="extract-content" Oct 04 04:14:52 crc kubenswrapper[4726]: E1004 04:14:52.253280 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="registry-server" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253286 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="registry-server" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253461 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a48f3207-0197-4361-958b-b7c491b3495c" containerName="registry-server" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.253475 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="dec26676-436d-4d9a-8337-0c9ac86c6168" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.254121 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.257021 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.257145 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.257844 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.261530 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.265453 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mjmvr"] Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.295261 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.295334 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6blgm\" (UniqueName: \"kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.295419 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.397781 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.398016 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.398085 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6blgm\" (UniqueName: \"kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.403344 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.404341 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.430162 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6blgm\" (UniqueName: \"kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm\") pod \"ssh-known-hosts-edpm-deployment-mjmvr\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:52 crc kubenswrapper[4726]: I1004 04:14:52.579715 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.239848 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mjmvr"] Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.363510 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.370452 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.409572 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.418277 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.418326 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.418368 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26dq7\" (UniqueName: \"kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.519603 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.519654 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.519696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26dq7\" (UniqueName: \"kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.521117 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.521810 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.541124 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26dq7\" (UniqueName: \"kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7\") pod \"redhat-marketplace-rp96n\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:53 crc kubenswrapper[4726]: I1004 04:14:53.708646 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:14:54 crc kubenswrapper[4726]: I1004 04:14:54.136675 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" event={"ID":"e0ad6b26-b887-46ea-9402-1e9bf3a77de9","Type":"ContainerStarted","Data":"0fa35a5503dbc4c23cab0034331320524ea73be99e2650a042849a8db3aa0fdd"} Oct 04 04:14:54 crc kubenswrapper[4726]: I1004 04:14:54.137798 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" event={"ID":"e0ad6b26-b887-46ea-9402-1e9bf3a77de9","Type":"ContainerStarted","Data":"1764136a506857a4b4e1a6777edaa50f6c2ed867ae643299935534249f87bd0c"} Oct 04 04:14:54 crc kubenswrapper[4726]: I1004 04:14:54.159640 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" podStartSLOduration=1.564918804 podStartE2EDuration="2.159622045s" podCreationTimestamp="2025-10-04 04:14:52 +0000 UTC" firstStartedPulling="2025-10-04 04:14:53.250593584 +0000 UTC m=+2067.425216837" lastFinishedPulling="2025-10-04 04:14:53.845296845 +0000 UTC m=+2068.019920078" observedRunningTime="2025-10-04 04:14:54.155899299 +0000 UTC m=+2068.330522532" watchObservedRunningTime="2025-10-04 04:14:54.159622045 +0000 UTC m=+2068.334245278" Oct 04 04:14:54 crc kubenswrapper[4726]: I1004 04:14:54.269476 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:14:55 crc kubenswrapper[4726]: I1004 04:14:55.150197 4726 generic.go:334] "Generic (PLEG): container finished" podID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerID="3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675" exitCode=0 Oct 04 04:14:55 crc kubenswrapper[4726]: I1004 04:14:55.150252 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerDied","Data":"3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675"} Oct 04 04:14:55 crc kubenswrapper[4726]: I1004 04:14:55.150607 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerStarted","Data":"f7a413b9e478d4413ae2434766dcca6eb09cef1750cd0953912db275c76c8584"} Oct 04 04:14:56 crc kubenswrapper[4726]: I1004 04:14:56.167049 4726 generic.go:334] "Generic (PLEG): container finished" podID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerID="0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11" exitCode=0 Oct 04 04:14:56 crc kubenswrapper[4726]: I1004 04:14:56.167091 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerDied","Data":"0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11"} Oct 04 04:14:56 crc kubenswrapper[4726]: I1004 04:14:56.914344 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:14:56 crc kubenswrapper[4726]: I1004 04:14:56.917447 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:56 crc kubenswrapper[4726]: I1004 04:14:56.938447 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.103891 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.104010 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrlxj\" (UniqueName: \"kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.104235 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.181024 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerStarted","Data":"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c"} Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.206477 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.206535 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrlxj\" (UniqueName: \"kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.206597 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.207100 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.207338 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.218780 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rp96n" podStartSLOduration=2.450358924 podStartE2EDuration="4.218761759s" podCreationTimestamp="2025-10-04 04:14:53 +0000 UTC" firstStartedPulling="2025-10-04 04:14:55.152885751 +0000 UTC m=+2069.327508964" lastFinishedPulling="2025-10-04 04:14:56.921288565 +0000 UTC m=+2071.095911799" observedRunningTime="2025-10-04 04:14:57.21261168 +0000 UTC m=+2071.387234893" watchObservedRunningTime="2025-10-04 04:14:57.218761759 +0000 UTC m=+2071.393384972" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.234884 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrlxj\" (UniqueName: \"kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj\") pod \"certified-operators-78lkt\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.248593 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:14:57 crc kubenswrapper[4726]: I1004 04:14:57.742720 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:14:57 crc kubenswrapper[4726]: W1004 04:14:57.752936 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf890f6b_8d44_407b_90d6_e42c9b74bba8.slice/crio-5df87fa08db6b33dd4cfab3615d7f3b77d337025562ff84999c27de5251fc14a WatchSource:0}: Error finding container 5df87fa08db6b33dd4cfab3615d7f3b77d337025562ff84999c27de5251fc14a: Status 404 returned error can't find the container with id 5df87fa08db6b33dd4cfab3615d7f3b77d337025562ff84999c27de5251fc14a Oct 04 04:14:58 crc kubenswrapper[4726]: I1004 04:14:58.191647 4726 generic.go:334] "Generic (PLEG): container finished" podID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerID="cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb" exitCode=0 Oct 04 04:14:58 crc kubenswrapper[4726]: I1004 04:14:58.191735 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerDied","Data":"cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb"} Oct 04 04:14:58 crc kubenswrapper[4726]: I1004 04:14:58.191790 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerStarted","Data":"5df87fa08db6b33dd4cfab3615d7f3b77d337025562ff84999c27de5251fc14a"} Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.315785 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.319660 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.328407 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.452798 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.452864 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdmr6\" (UniqueName: \"kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.453071 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.554688 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdmr6\" (UniqueName: \"kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.554809 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.554922 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.555287 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.555357 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.579519 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdmr6\" (UniqueName: \"kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6\") pod \"redhat-operators-mxthd\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.646617 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:14:59 crc kubenswrapper[4726]: I1004 04:14:59.919244 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.136976 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9"] Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.138787 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.145817 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.146256 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.150481 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9"] Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.211556 4726 generic.go:334] "Generic (PLEG): container finished" podID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerID="3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5" exitCode=0 Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.211806 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerDied","Data":"3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5"} Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.211835 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerStarted","Data":"3d0114ff2e941646d21d2496000dc35fb518b6bf4d4312a4987bb79110e884c6"} Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.213739 4726 generic.go:334] "Generic (PLEG): container finished" podID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerID="f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa" exitCode=0 Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.213764 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerDied","Data":"f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa"} Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.267303 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxk88\" (UniqueName: \"kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.267394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.267483 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.368996 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.369130 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxk88\" (UniqueName: \"kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.369455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.370274 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.380196 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.395331 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxk88\" (UniqueName: \"kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88\") pod \"collect-profiles-29325855-7rjw9\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.468521 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:00 crc kubenswrapper[4726]: I1004 04:15:00.978120 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9"] Oct 04 04:15:01 crc kubenswrapper[4726]: I1004 04:15:01.227004 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" event={"ID":"3910d065-7f83-4b54-90e8-b78a919b7084","Type":"ContainerStarted","Data":"d16878d5c4a54d8ac72eab6db0bf520f36b15a2d664a399da0212f4cdc139e2c"} Oct 04 04:15:01 crc kubenswrapper[4726]: I1004 04:15:01.230834 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerStarted","Data":"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0"} Oct 04 04:15:01 crc kubenswrapper[4726]: I1004 04:15:01.259765 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-78lkt" podStartSLOduration=2.737389228 podStartE2EDuration="5.259751483s" podCreationTimestamp="2025-10-04 04:14:56 +0000 UTC" firstStartedPulling="2025-10-04 04:14:58.194436799 +0000 UTC m=+2072.369060052" lastFinishedPulling="2025-10-04 04:15:00.716799094 +0000 UTC m=+2074.891422307" observedRunningTime="2025-10-04 04:15:01.252532446 +0000 UTC m=+2075.427155659" watchObservedRunningTime="2025-10-04 04:15:01.259751483 +0000 UTC m=+2075.434374696" Oct 04 04:15:02 crc kubenswrapper[4726]: I1004 04:15:02.251409 4726 generic.go:334] "Generic (PLEG): container finished" podID="e0ad6b26-b887-46ea-9402-1e9bf3a77de9" containerID="0fa35a5503dbc4c23cab0034331320524ea73be99e2650a042849a8db3aa0fdd" exitCode=0 Oct 04 04:15:02 crc kubenswrapper[4726]: I1004 04:15:02.251591 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" event={"ID":"e0ad6b26-b887-46ea-9402-1e9bf3a77de9","Type":"ContainerDied","Data":"0fa35a5503dbc4c23cab0034331320524ea73be99e2650a042849a8db3aa0fdd"} Oct 04 04:15:02 crc kubenswrapper[4726]: I1004 04:15:02.254664 4726 generic.go:334] "Generic (PLEG): container finished" podID="3910d065-7f83-4b54-90e8-b78a919b7084" containerID="32079598df637351418db54f1c82a234ec0d465b00fb937566fbcbd8f581ccfd" exitCode=0 Oct 04 04:15:02 crc kubenswrapper[4726]: I1004 04:15:02.254754 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" event={"ID":"3910d065-7f83-4b54-90e8-b78a919b7084","Type":"ContainerDied","Data":"32079598df637351418db54f1c82a234ec0d465b00fb937566fbcbd8f581ccfd"} Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.269413 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerStarted","Data":"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37"} Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.709009 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.709288 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.757955 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.775262 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.782891 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.847914 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxk88\" (UniqueName: \"kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88\") pod \"3910d065-7f83-4b54-90e8-b78a919b7084\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.848197 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam\") pod \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.848300 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume\") pod \"3910d065-7f83-4b54-90e8-b78a919b7084\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.848437 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume\") pod \"3910d065-7f83-4b54-90e8-b78a919b7084\" (UID: \"3910d065-7f83-4b54-90e8-b78a919b7084\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.848579 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6blgm\" (UniqueName: \"kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm\") pod \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.848871 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0\") pod \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\" (UID: \"e0ad6b26-b887-46ea-9402-1e9bf3a77de9\") " Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.856227 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume" (OuterVolumeSpecName: "config-volume") pod "3910d065-7f83-4b54-90e8-b78a919b7084" (UID: "3910d065-7f83-4b54-90e8-b78a919b7084"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.857174 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3910d065-7f83-4b54-90e8-b78a919b7084" (UID: "3910d065-7f83-4b54-90e8-b78a919b7084"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.862914 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm" (OuterVolumeSpecName: "kube-api-access-6blgm") pod "e0ad6b26-b887-46ea-9402-1e9bf3a77de9" (UID: "e0ad6b26-b887-46ea-9402-1e9bf3a77de9"). InnerVolumeSpecName "kube-api-access-6blgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.863150 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88" (OuterVolumeSpecName: "kube-api-access-cxk88") pod "3910d065-7f83-4b54-90e8-b78a919b7084" (UID: "3910d065-7f83-4b54-90e8-b78a919b7084"). InnerVolumeSpecName "kube-api-access-cxk88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.887267 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e0ad6b26-b887-46ea-9402-1e9bf3a77de9" (UID: "e0ad6b26-b887-46ea-9402-1e9bf3a77de9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.891340 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e0ad6b26-b887-46ea-9402-1e9bf3a77de9" (UID: "e0ad6b26-b887-46ea-9402-1e9bf3a77de9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.952766 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxk88\" (UniqueName: \"kubernetes.io/projected/3910d065-7f83-4b54-90e8-b78a919b7084-kube-api-access-cxk88\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.953082 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.953258 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3910d065-7f83-4b54-90e8-b78a919b7084-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.953378 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3910d065-7f83-4b54-90e8-b78a919b7084-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.953489 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6blgm\" (UniqueName: \"kubernetes.io/projected/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-kube-api-access-6blgm\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:03 crc kubenswrapper[4726]: I1004 04:15:03.953595 4726 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e0ad6b26-b887-46ea-9402-1e9bf3a77de9-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.188254 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.188355 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.188423 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.189349 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.189470 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb" gracePeriod=600 Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.283012 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.282977 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325855-7rjw9" event={"ID":"3910d065-7f83-4b54-90e8-b78a919b7084","Type":"ContainerDied","Data":"d16878d5c4a54d8ac72eab6db0bf520f36b15a2d664a399da0212f4cdc139e2c"} Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.284372 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d16878d5c4a54d8ac72eab6db0bf520f36b15a2d664a399da0212f4cdc139e2c" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.291406 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.291414 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mjmvr" event={"ID":"e0ad6b26-b887-46ea-9402-1e9bf3a77de9","Type":"ContainerDied","Data":"1764136a506857a4b4e1a6777edaa50f6c2ed867ae643299935534249f87bd0c"} Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.291488 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1764136a506857a4b4e1a6777edaa50f6c2ed867ae643299935534249f87bd0c" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.362330 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8"] Oct 04 04:15:04 crc kubenswrapper[4726]: E1004 04:15:04.363075 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3910d065-7f83-4b54-90e8-b78a919b7084" containerName="collect-profiles" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.363099 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3910d065-7f83-4b54-90e8-b78a919b7084" containerName="collect-profiles" Oct 04 04:15:04 crc kubenswrapper[4726]: E1004 04:15:04.363162 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ad6b26-b887-46ea-9402-1e9bf3a77de9" containerName="ssh-known-hosts-edpm-deployment" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.363172 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ad6b26-b887-46ea-9402-1e9bf3a77de9" containerName="ssh-known-hosts-edpm-deployment" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.363451 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3910d065-7f83-4b54-90e8-b78a919b7084" containerName="collect-profiles" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.363489 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ad6b26-b887-46ea-9402-1e9bf3a77de9" containerName="ssh-known-hosts-edpm-deployment" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.364149 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.364243 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.411324 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.411403 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.411590 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.411678 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.424554 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8"] Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.463575 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.463641 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.463713 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdvmf\" (UniqueName: \"kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.565666 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.565729 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.565806 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdvmf\" (UniqueName: \"kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.571726 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.572317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.581088 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdvmf\" (UniqueName: \"kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-786b8\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.730755 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.894027 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq"] Oct 04 04:15:04 crc kubenswrapper[4726]: I1004 04:15:04.902557 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325810-gkptq"] Oct 04 04:15:05 crc kubenswrapper[4726]: I1004 04:15:05.305588 4726 generic.go:334] "Generic (PLEG): container finished" podID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerID="59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37" exitCode=0 Oct 04 04:15:05 crc kubenswrapper[4726]: I1004 04:15:05.306690 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerDied","Data":"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37"} Oct 04 04:15:05 crc kubenswrapper[4726]: I1004 04:15:05.642868 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8"] Oct 04 04:15:05 crc kubenswrapper[4726]: W1004 04:15:05.651006 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9465de2d_b388_410c_90da_31beba8b5dbe.slice/crio-77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f WatchSource:0}: Error finding container 77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f: Status 404 returned error can't find the container with id 77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f Oct 04 04:15:06 crc kubenswrapper[4726]: I1004 04:15:06.107545 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:15:06 crc kubenswrapper[4726]: I1004 04:15:06.315577 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" event={"ID":"9465de2d-b388-410c-90da-31beba8b5dbe","Type":"ContainerStarted","Data":"77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f"} Oct 04 04:15:06 crc kubenswrapper[4726]: I1004 04:15:06.315744 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rp96n" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="registry-server" containerID="cri-o://37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c" gracePeriod=2 Oct 04 04:15:06 crc kubenswrapper[4726]: I1004 04:15:06.529218 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d249e87-e695-4058-8b58-a068fcd85857" path="/var/lib/kubelet/pods/1d249e87-e695-4058-8b58-a068fcd85857/volumes" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.249494 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.249561 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.257855 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.310638 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.330158 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities\") pod \"82c7f048-9c11-48dd-a450-53c5c5b765db\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.330658 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26dq7\" (UniqueName: \"kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7\") pod \"82c7f048-9c11-48dd-a450-53c5c5b765db\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.330778 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content\") pod \"82c7f048-9c11-48dd-a450-53c5c5b765db\" (UID: \"82c7f048-9c11-48dd-a450-53c5c5b765db\") " Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.331384 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" event={"ID":"9465de2d-b388-410c-90da-31beba8b5dbe","Type":"ContainerStarted","Data":"09b95c34c250a563f7c3ff594a53ea66c5317734988f8e8c30d6312a68b5360b"} Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.331530 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities" (OuterVolumeSpecName: "utilities") pod "82c7f048-9c11-48dd-a450-53c5c5b765db" (UID: "82c7f048-9c11-48dd-a450-53c5c5b765db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.339278 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb" exitCode=0 Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.339351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb"} Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.339385 4726 scope.go:117] "RemoveContainer" containerID="abed132d638bf9b22c93abb9dcc535aeea603731ab2397ef7116f477cd979fe9" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.342680 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7" (OuterVolumeSpecName: "kube-api-access-26dq7") pod "82c7f048-9c11-48dd-a450-53c5c5b765db" (UID: "82c7f048-9c11-48dd-a450-53c5c5b765db"). InnerVolumeSpecName "kube-api-access-26dq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.343673 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerStarted","Data":"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e"} Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.346977 4726 generic.go:334] "Generic (PLEG): container finished" podID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerID="37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c" exitCode=0 Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.347028 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rp96n" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.347097 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerDied","Data":"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c"} Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.347200 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rp96n" event={"ID":"82c7f048-9c11-48dd-a450-53c5c5b765db","Type":"ContainerDied","Data":"f7a413b9e478d4413ae2434766dcca6eb09cef1750cd0953912db275c76c8584"} Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.361846 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82c7f048-9c11-48dd-a450-53c5c5b765db" (UID: "82c7f048-9c11-48dd-a450-53c5c5b765db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.362529 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" podStartSLOduration=2.371477091 podStartE2EDuration="3.362505848s" podCreationTimestamp="2025-10-04 04:15:04 +0000 UTC" firstStartedPulling="2025-10-04 04:15:05.653569176 +0000 UTC m=+2079.828192389" lastFinishedPulling="2025-10-04 04:15:06.644597933 +0000 UTC m=+2080.819221146" observedRunningTime="2025-10-04 04:15:07.35177026 +0000 UTC m=+2081.526393483" watchObservedRunningTime="2025-10-04 04:15:07.362505848 +0000 UTC m=+2081.537129061" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.375505 4726 scope.go:117] "RemoveContainer" containerID="37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.378661 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mxthd" podStartSLOduration=3.105360039 podStartE2EDuration="8.378643587s" podCreationTimestamp="2025-10-04 04:14:59 +0000 UTC" firstStartedPulling="2025-10-04 04:15:01.232302111 +0000 UTC m=+2075.406925324" lastFinishedPulling="2025-10-04 04:15:06.505585619 +0000 UTC m=+2080.680208872" observedRunningTime="2025-10-04 04:15:07.371966803 +0000 UTC m=+2081.546590016" watchObservedRunningTime="2025-10-04 04:15:07.378643587 +0000 UTC m=+2081.553266800" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.397515 4726 scope.go:117] "RemoveContainer" containerID="0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.400074 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.417219 4726 scope.go:117] "RemoveContainer" containerID="3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.434166 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.434209 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26dq7\" (UniqueName: \"kubernetes.io/projected/82c7f048-9c11-48dd-a450-53c5c5b765db-kube-api-access-26dq7\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.434219 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c7f048-9c11-48dd-a450-53c5c5b765db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.435428 4726 scope.go:117] "RemoveContainer" containerID="37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c" Oct 04 04:15:07 crc kubenswrapper[4726]: E1004 04:15:07.435762 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c\": container with ID starting with 37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c not found: ID does not exist" containerID="37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.435791 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c"} err="failed to get container status \"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c\": rpc error: code = NotFound desc = could not find container \"37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c\": container with ID starting with 37b573b96e8c8c2fa5dfab532bfb57cf06fe26ec76c049bf269cd58d30b59e4c not found: ID does not exist" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.435809 4726 scope.go:117] "RemoveContainer" containerID="0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11" Oct 04 04:15:07 crc kubenswrapper[4726]: E1004 04:15:07.436047 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11\": container with ID starting with 0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11 not found: ID does not exist" containerID="0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.436093 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11"} err="failed to get container status \"0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11\": rpc error: code = NotFound desc = could not find container \"0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11\": container with ID starting with 0ceb6b2764aa9fb0e799d1ad1834939d9eb36f0bdc4dee3f67d823ef2978df11 not found: ID does not exist" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.436180 4726 scope.go:117] "RemoveContainer" containerID="3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675" Oct 04 04:15:07 crc kubenswrapper[4726]: E1004 04:15:07.436529 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675\": container with ID starting with 3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675 not found: ID does not exist" containerID="3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.436553 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675"} err="failed to get container status \"3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675\": rpc error: code = NotFound desc = could not find container \"3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675\": container with ID starting with 3d89a7dddaa7217be7e1a314c86f3fab5bbfce5b5cb30f5174eed017cab07675 not found: ID does not exist" Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.757173 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:15:07 crc kubenswrapper[4726]: I1004 04:15:07.774546 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rp96n"] Oct 04 04:15:08 crc kubenswrapper[4726]: I1004 04:15:08.364092 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7"} Oct 04 04:15:08 crc kubenswrapper[4726]: I1004 04:15:08.538923 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" path="/var/lib/kubelet/pods/82c7f048-9c11-48dd-a450-53c5c5b765db/volumes" Oct 04 04:15:09 crc kubenswrapper[4726]: I1004 04:15:09.647421 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:09 crc kubenswrapper[4726]: I1004 04:15:09.647893 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:10 crc kubenswrapper[4726]: I1004 04:15:10.501583 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:15:10 crc kubenswrapper[4726]: I1004 04:15:10.502308 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-78lkt" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="registry-server" containerID="cri-o://51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0" gracePeriod=2 Oct 04 04:15:10 crc kubenswrapper[4726]: I1004 04:15:10.718760 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mxthd" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="registry-server" probeResult="failure" output=< Oct 04 04:15:10 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:15:10 crc kubenswrapper[4726]: > Oct 04 04:15:10 crc kubenswrapper[4726]: I1004 04:15:10.943536 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.119458 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities\") pod \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.119651 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrlxj\" (UniqueName: \"kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj\") pod \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.119688 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content\") pod \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\" (UID: \"cf890f6b-8d44-407b-90d6-e42c9b74bba8\") " Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.120467 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities" (OuterVolumeSpecName: "utilities") pod "cf890f6b-8d44-407b-90d6-e42c9b74bba8" (UID: "cf890f6b-8d44-407b-90d6-e42c9b74bba8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.127951 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj" (OuterVolumeSpecName: "kube-api-access-mrlxj") pod "cf890f6b-8d44-407b-90d6-e42c9b74bba8" (UID: "cf890f6b-8d44-407b-90d6-e42c9b74bba8"). InnerVolumeSpecName "kube-api-access-mrlxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.173309 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf890f6b-8d44-407b-90d6-e42c9b74bba8" (UID: "cf890f6b-8d44-407b-90d6-e42c9b74bba8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.229829 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrlxj\" (UniqueName: \"kubernetes.io/projected/cf890f6b-8d44-407b-90d6-e42c9b74bba8-kube-api-access-mrlxj\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.229877 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.229897 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf890f6b-8d44-407b-90d6-e42c9b74bba8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.433352 4726 generic.go:334] "Generic (PLEG): container finished" podID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerID="51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0" exitCode=0 Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.433410 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerDied","Data":"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0"} Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.433445 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78lkt" event={"ID":"cf890f6b-8d44-407b-90d6-e42c9b74bba8","Type":"ContainerDied","Data":"5df87fa08db6b33dd4cfab3615d7f3b77d337025562ff84999c27de5251fc14a"} Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.433472 4726 scope.go:117] "RemoveContainer" containerID="51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.433485 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78lkt" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.467709 4726 scope.go:117] "RemoveContainer" containerID="f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.492030 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.502235 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-78lkt"] Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.516277 4726 scope.go:117] "RemoveContainer" containerID="cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.548340 4726 scope.go:117] "RemoveContainer" containerID="51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0" Oct 04 04:15:11 crc kubenswrapper[4726]: E1004 04:15:11.548925 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0\": container with ID starting with 51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0 not found: ID does not exist" containerID="51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.548966 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0"} err="failed to get container status \"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0\": rpc error: code = NotFound desc = could not find container \"51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0\": container with ID starting with 51a49d47ccc9e525f9bb6e25bbe9377df0a43d730bdd6c062454d8d93156d8b0 not found: ID does not exist" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.548996 4726 scope.go:117] "RemoveContainer" containerID="f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa" Oct 04 04:15:11 crc kubenswrapper[4726]: E1004 04:15:11.549478 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa\": container with ID starting with f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa not found: ID does not exist" containerID="f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.549544 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa"} err="failed to get container status \"f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa\": rpc error: code = NotFound desc = could not find container \"f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa\": container with ID starting with f062ea01db9a30efb8bee016d968f40d92887290e1589c2193f0af9cc2e826fa not found: ID does not exist" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.549586 4726 scope.go:117] "RemoveContainer" containerID="cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb" Oct 04 04:15:11 crc kubenswrapper[4726]: E1004 04:15:11.549954 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb\": container with ID starting with cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb not found: ID does not exist" containerID="cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb" Oct 04 04:15:11 crc kubenswrapper[4726]: I1004 04:15:11.549980 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb"} err="failed to get container status \"cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb\": rpc error: code = NotFound desc = could not find container \"cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb\": container with ID starting with cc0a35301b672d2f6cb916a67fbe8b5dc10137e5d11f0f0de1a0ecc48efc97fb not found: ID does not exist" Oct 04 04:15:12 crc kubenswrapper[4726]: I1004 04:15:12.524273 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" path="/var/lib/kubelet/pods/cf890f6b-8d44-407b-90d6-e42c9b74bba8/volumes" Oct 04 04:15:16 crc kubenswrapper[4726]: I1004 04:15:16.525293 4726 generic.go:334] "Generic (PLEG): container finished" podID="9465de2d-b388-410c-90da-31beba8b5dbe" containerID="09b95c34c250a563f7c3ff594a53ea66c5317734988f8e8c30d6312a68b5360b" exitCode=0 Oct 04 04:15:16 crc kubenswrapper[4726]: I1004 04:15:16.525372 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" event={"ID":"9465de2d-b388-410c-90da-31beba8b5dbe","Type":"ContainerDied","Data":"09b95c34c250a563f7c3ff594a53ea66c5317734988f8e8c30d6312a68b5360b"} Oct 04 04:15:17 crc kubenswrapper[4726]: I1004 04:15:17.955725 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.138644 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory\") pod \"9465de2d-b388-410c-90da-31beba8b5dbe\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.139444 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key\") pod \"9465de2d-b388-410c-90da-31beba8b5dbe\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.139698 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdvmf\" (UniqueName: \"kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf\") pod \"9465de2d-b388-410c-90da-31beba8b5dbe\" (UID: \"9465de2d-b388-410c-90da-31beba8b5dbe\") " Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.164083 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf" (OuterVolumeSpecName: "kube-api-access-mdvmf") pod "9465de2d-b388-410c-90da-31beba8b5dbe" (UID: "9465de2d-b388-410c-90da-31beba8b5dbe"). InnerVolumeSpecName "kube-api-access-mdvmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.182354 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory" (OuterVolumeSpecName: "inventory") pod "9465de2d-b388-410c-90da-31beba8b5dbe" (UID: "9465de2d-b388-410c-90da-31beba8b5dbe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.188171 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9465de2d-b388-410c-90da-31beba8b5dbe" (UID: "9465de2d-b388-410c-90da-31beba8b5dbe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.242198 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdvmf\" (UniqueName: \"kubernetes.io/projected/9465de2d-b388-410c-90da-31beba8b5dbe-kube-api-access-mdvmf\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.242231 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.242240 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9465de2d-b388-410c-90da-31beba8b5dbe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.557702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" event={"ID":"9465de2d-b388-410c-90da-31beba8b5dbe","Type":"ContainerDied","Data":"77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f"} Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.557813 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77799b1caed7e42e7efb7bb3aedb324e7e43e4466799036651b6b42b8dad2c8f" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.557964 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-786b8" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.647256 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969"] Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.647912 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.647934 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.647976 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="extract-utilities" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.647986 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="extract-utilities" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.648007 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="extract-utilities" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648033 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="extract-utilities" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.648056 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9465de2d-b388-410c-90da-31beba8b5dbe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648065 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="9465de2d-b388-410c-90da-31beba8b5dbe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.648091 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648099 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.648143 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="extract-content" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648151 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="extract-content" Oct 04 04:15:18 crc kubenswrapper[4726]: E1004 04:15:18.648167 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="extract-content" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648175 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="extract-content" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648430 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf890f6b-8d44-407b-90d6-e42c9b74bba8" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648447 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="9465de2d-b388-410c-90da-31beba8b5dbe" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.648504 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c7f048-9c11-48dd-a450-53c5c5b765db" containerName="registry-server" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.649280 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.658460 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.658852 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.659200 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.659397 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.685474 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969"] Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.754695 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.754777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.754924 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4pr9\" (UniqueName: \"kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.857015 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4pr9\" (UniqueName: \"kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.857163 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.857218 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.862087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.870298 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.874640 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4pr9\" (UniqueName: \"kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-w8969\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:18 crc kubenswrapper[4726]: I1004 04:15:18.979012 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:19 crc kubenswrapper[4726]: W1004 04:15:19.561649 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c4effe5_c8f2_4f82_b5b0_3052858642f1.slice/crio-a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497 WatchSource:0}: Error finding container a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497: Status 404 returned error can't find the container with id a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497 Oct 04 04:15:19 crc kubenswrapper[4726]: I1004 04:15:19.563217 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969"] Oct 04 04:15:19 crc kubenswrapper[4726]: I1004 04:15:19.731540 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:19 crc kubenswrapper[4726]: I1004 04:15:19.783610 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:19 crc kubenswrapper[4726]: I1004 04:15:19.968262 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:15:20 crc kubenswrapper[4726]: I1004 04:15:20.593550 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" event={"ID":"0c4effe5-c8f2-4f82-b5b0-3052858642f1","Type":"ContainerStarted","Data":"ce8946a3a3da518f9c8647ebc538a68ea96dcec5dc143a90d0aec93b4f8e4b61"} Oct 04 04:15:20 crc kubenswrapper[4726]: I1004 04:15:20.595774 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" event={"ID":"0c4effe5-c8f2-4f82-b5b0-3052858642f1","Type":"ContainerStarted","Data":"a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497"} Oct 04 04:15:20 crc kubenswrapper[4726]: I1004 04:15:20.627405 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" podStartSLOduration=2.203523296 podStartE2EDuration="2.627388317s" podCreationTimestamp="2025-10-04 04:15:18 +0000 UTC" firstStartedPulling="2025-10-04 04:15:19.571706293 +0000 UTC m=+2093.746329506" lastFinishedPulling="2025-10-04 04:15:19.995571324 +0000 UTC m=+2094.170194527" observedRunningTime="2025-10-04 04:15:20.62671216 +0000 UTC m=+2094.801335403" watchObservedRunningTime="2025-10-04 04:15:20.627388317 +0000 UTC m=+2094.802011530" Oct 04 04:15:21 crc kubenswrapper[4726]: I1004 04:15:21.603198 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mxthd" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="registry-server" containerID="cri-o://64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e" gracePeriod=2 Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.058527 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.224746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities\") pod \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.224926 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content\") pod \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.224989 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdmr6\" (UniqueName: \"kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6\") pod \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\" (UID: \"adc34fc7-beda-4cf0-9e75-5b9c767f64c7\") " Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.225903 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities" (OuterVolumeSpecName: "utilities") pod "adc34fc7-beda-4cf0-9e75-5b9c767f64c7" (UID: "adc34fc7-beda-4cf0-9e75-5b9c767f64c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.234917 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6" (OuterVolumeSpecName: "kube-api-access-sdmr6") pod "adc34fc7-beda-4cf0-9e75-5b9c767f64c7" (UID: "adc34fc7-beda-4cf0-9e75-5b9c767f64c7"). InnerVolumeSpecName "kube-api-access-sdmr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.327831 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.327873 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdmr6\" (UniqueName: \"kubernetes.io/projected/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-kube-api-access-sdmr6\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.338160 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adc34fc7-beda-4cf0-9e75-5b9c767f64c7" (UID: "adc34fc7-beda-4cf0-9e75-5b9c767f64c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.430619 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adc34fc7-beda-4cf0-9e75-5b9c767f64c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.620240 4726 generic.go:334] "Generic (PLEG): container finished" podID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerID="64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e" exitCode=0 Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.620306 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerDied","Data":"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e"} Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.620395 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mxthd" event={"ID":"adc34fc7-beda-4cf0-9e75-5b9c767f64c7","Type":"ContainerDied","Data":"3d0114ff2e941646d21d2496000dc35fb518b6bf4d4312a4987bb79110e884c6"} Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.620433 4726 scope.go:117] "RemoveContainer" containerID="64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.620715 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mxthd" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.665965 4726 scope.go:117] "RemoveContainer" containerID="59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.682916 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.698428 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mxthd"] Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.706387 4726 scope.go:117] "RemoveContainer" containerID="3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.763122 4726 scope.go:117] "RemoveContainer" containerID="64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e" Oct 04 04:15:22 crc kubenswrapper[4726]: E1004 04:15:22.763642 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e\": container with ID starting with 64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e not found: ID does not exist" containerID="64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.763688 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e"} err="failed to get container status \"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e\": rpc error: code = NotFound desc = could not find container \"64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e\": container with ID starting with 64a1313d0219e8af4b7c211ae689717424b33de4ba74ee26b377242032b6465e not found: ID does not exist" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.763715 4726 scope.go:117] "RemoveContainer" containerID="59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37" Oct 04 04:15:22 crc kubenswrapper[4726]: E1004 04:15:22.764038 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37\": container with ID starting with 59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37 not found: ID does not exist" containerID="59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.764081 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37"} err="failed to get container status \"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37\": rpc error: code = NotFound desc = could not find container \"59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37\": container with ID starting with 59cb7ceb1a95011015daead6b43bdc5f6246599f2105b6215203f058edf51c37 not found: ID does not exist" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.764126 4726 scope.go:117] "RemoveContainer" containerID="3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5" Oct 04 04:15:22 crc kubenswrapper[4726]: E1004 04:15:22.764453 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5\": container with ID starting with 3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5 not found: ID does not exist" containerID="3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5" Oct 04 04:15:22 crc kubenswrapper[4726]: I1004 04:15:22.764492 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5"} err="failed to get container status \"3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5\": rpc error: code = NotFound desc = could not find container \"3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5\": container with ID starting with 3915b03e28b41d2c12f308c1c6b4132c551087fbb631357c33fb29efa95bcab5 not found: ID does not exist" Oct 04 04:15:24 crc kubenswrapper[4726]: I1004 04:15:24.521246 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" path="/var/lib/kubelet/pods/adc34fc7-beda-4cf0-9e75-5b9c767f64c7/volumes" Oct 04 04:15:30 crc kubenswrapper[4726]: I1004 04:15:30.728091 4726 generic.go:334] "Generic (PLEG): container finished" podID="0c4effe5-c8f2-4f82-b5b0-3052858642f1" containerID="ce8946a3a3da518f9c8647ebc538a68ea96dcec5dc143a90d0aec93b4f8e4b61" exitCode=0 Oct 04 04:15:30 crc kubenswrapper[4726]: I1004 04:15:30.728267 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" event={"ID":"0c4effe5-c8f2-4f82-b5b0-3052858642f1","Type":"ContainerDied","Data":"ce8946a3a3da518f9c8647ebc538a68ea96dcec5dc143a90d0aec93b4f8e4b61"} Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.211926 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.340996 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4pr9\" (UniqueName: \"kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9\") pod \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.341579 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key\") pod \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.341706 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory\") pod \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\" (UID: \"0c4effe5-c8f2-4f82-b5b0-3052858642f1\") " Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.348440 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9" (OuterVolumeSpecName: "kube-api-access-d4pr9") pod "0c4effe5-c8f2-4f82-b5b0-3052858642f1" (UID: "0c4effe5-c8f2-4f82-b5b0-3052858642f1"). InnerVolumeSpecName "kube-api-access-d4pr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.369838 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory" (OuterVolumeSpecName: "inventory") pod "0c4effe5-c8f2-4f82-b5b0-3052858642f1" (UID: "0c4effe5-c8f2-4f82-b5b0-3052858642f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.376208 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0c4effe5-c8f2-4f82-b5b0-3052858642f1" (UID: "0c4effe5-c8f2-4f82-b5b0-3052858642f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.444191 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.444229 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4pr9\" (UniqueName: \"kubernetes.io/projected/0c4effe5-c8f2-4f82-b5b0-3052858642f1-kube-api-access-d4pr9\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.444242 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0c4effe5-c8f2-4f82-b5b0-3052858642f1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.752910 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" event={"ID":"0c4effe5-c8f2-4f82-b5b0-3052858642f1","Type":"ContainerDied","Data":"a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497"} Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.752971 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b55feaaf610b6ebb113e948853697c7e0fe4fe1aa7cfc1a2d4dd65c4039497" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.753149 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-w8969" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.873477 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft"] Oct 04 04:15:32 crc kubenswrapper[4726]: E1004 04:15:32.873900 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="extract-utilities" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.873924 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="extract-utilities" Oct 04 04:15:32 crc kubenswrapper[4726]: E1004 04:15:32.873943 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="registry-server" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.873951 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="registry-server" Oct 04 04:15:32 crc kubenswrapper[4726]: E1004 04:15:32.873969 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c4effe5-c8f2-4f82-b5b0-3052858642f1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.873980 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c4effe5-c8f2-4f82-b5b0-3052858642f1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:32 crc kubenswrapper[4726]: E1004 04:15:32.873990 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="extract-content" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.873998 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="extract-content" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.874233 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c4effe5-c8f2-4f82-b5b0-3052858642f1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.874260 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc34fc7-beda-4cf0-9e75-5b9c767f64c7" containerName="registry-server" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.874965 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.878347 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.878746 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.879838 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.879932 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.879955 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.880608 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.880714 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.880982 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:15:32 crc kubenswrapper[4726]: I1004 04:15:32.909486 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft"] Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065467 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065544 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065668 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065704 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065735 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065761 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tct8\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065949 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.065978 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.066021 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.066077 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.066168 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.066267 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.066332 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168172 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168245 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168302 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168337 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168393 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168434 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168467 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168505 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168558 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168582 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168605 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168629 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tct8\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168663 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.168707 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.174962 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.176190 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.176744 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.178943 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.179030 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.179681 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.179791 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.179977 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.180193 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.181203 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.182413 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.184601 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.190855 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.199670 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tct8\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nzrft\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.248265 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.651730 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft"] Oct 04 04:15:33 crc kubenswrapper[4726]: I1004 04:15:33.767268 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" event={"ID":"70435c0c-e859-49bd-9e2e-624f66c3be54","Type":"ContainerStarted","Data":"b213a1b6eb2f41db13e1fbba926ea6504ded7981931f1c02b4ce8aab333c88ca"} Oct 04 04:15:34 crc kubenswrapper[4726]: I1004 04:15:34.776788 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" event={"ID":"70435c0c-e859-49bd-9e2e-624f66c3be54","Type":"ContainerStarted","Data":"d2514a87d53dbe58693e348384ef4ddd7a7b9372b4b8398f4f756057deb88129"} Oct 04 04:15:34 crc kubenswrapper[4726]: I1004 04:15:34.809393 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" podStartSLOduration=2.361668988 podStartE2EDuration="2.809375288s" podCreationTimestamp="2025-10-04 04:15:32 +0000 UTC" firstStartedPulling="2025-10-04 04:15:33.657431937 +0000 UTC m=+2107.832055160" lastFinishedPulling="2025-10-04 04:15:34.105138207 +0000 UTC m=+2108.279761460" observedRunningTime="2025-10-04 04:15:34.801235907 +0000 UTC m=+2108.975859120" watchObservedRunningTime="2025-10-04 04:15:34.809375288 +0000 UTC m=+2108.983998501" Oct 04 04:15:37 crc kubenswrapper[4726]: I1004 04:15:37.586850 4726 scope.go:117] "RemoveContainer" containerID="edfad6528dff4afdcd90b4b127589d14fd1164b5f9fdb0db13e19971f609cd05" Oct 04 04:16:16 crc kubenswrapper[4726]: I1004 04:16:16.264533 4726 generic.go:334] "Generic (PLEG): container finished" podID="70435c0c-e859-49bd-9e2e-624f66c3be54" containerID="d2514a87d53dbe58693e348384ef4ddd7a7b9372b4b8398f4f756057deb88129" exitCode=0 Oct 04 04:16:16 crc kubenswrapper[4726]: I1004 04:16:16.265206 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" event={"ID":"70435c0c-e859-49bd-9e2e-624f66c3be54","Type":"ContainerDied","Data":"d2514a87d53dbe58693e348384ef4ddd7a7b9372b4b8398f4f756057deb88129"} Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.760625 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.869528 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.869601 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.869731 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tct8\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.869752 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870746 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870789 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870842 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870876 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870914 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.870987 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.871018 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.871058 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.871083 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.871252 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle\") pod \"70435c0c-e859-49bd-9e2e-624f66c3be54\" (UID: \"70435c0c-e859-49bd-9e2e-624f66c3be54\") " Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.876028 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.876596 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8" (OuterVolumeSpecName: "kube-api-access-8tct8") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "kube-api-access-8tct8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.877045 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.877149 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.877510 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.877914 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.880193 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.880222 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.880375 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.880779 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.881609 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.907237 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.908758 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory" (OuterVolumeSpecName: "inventory") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.929628 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70435c0c-e859-49bd-9e2e-624f66c3be54" (UID: "70435c0c-e859-49bd-9e2e-624f66c3be54"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973743 4726 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973791 4726 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973806 4726 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973820 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973834 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973847 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tct8\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-kube-api-access-8tct8\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973864 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973877 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973889 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973904 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973917 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973929 4726 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70435c0c-e859-49bd-9e2e-624f66c3be54-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973943 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:17 crc kubenswrapper[4726]: I1004 04:16:17.973955 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/70435c0c-e859-49bd-9e2e-624f66c3be54-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.289954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" event={"ID":"70435c0c-e859-49bd-9e2e-624f66c3be54","Type":"ContainerDied","Data":"b213a1b6eb2f41db13e1fbba926ea6504ded7981931f1c02b4ce8aab333c88ca"} Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.290309 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b213a1b6eb2f41db13e1fbba926ea6504ded7981931f1c02b4ce8aab333c88ca" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.290042 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nzrft" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.405381 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp"] Oct 04 04:16:18 crc kubenswrapper[4726]: E1004 04:16:18.405800 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70435c0c-e859-49bd-9e2e-624f66c3be54" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.405818 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="70435c0c-e859-49bd-9e2e-624f66c3be54" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.405997 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="70435c0c-e859-49bd-9e2e-624f66c3be54" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.406622 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.409680 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.410728 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.413875 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.414674 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.414703 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.425913 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp"] Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.584192 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.584617 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5v2l\" (UniqueName: \"kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.584898 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.585156 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.585547 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.687756 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.687906 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.688084 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5v2l\" (UniqueName: \"kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.688192 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.688275 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.689415 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.693284 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.693673 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.693985 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.708816 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5v2l\" (UniqueName: \"kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-jfxbp\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:18 crc kubenswrapper[4726]: I1004 04:16:18.727772 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:16:19 crc kubenswrapper[4726]: I1004 04:16:19.083527 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp"] Oct 04 04:16:19 crc kubenswrapper[4726]: W1004 04:16:19.089836 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1087fbc_edc3_47d2_8229_fb1654ce31eb.slice/crio-561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7 WatchSource:0}: Error finding container 561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7: Status 404 returned error can't find the container with id 561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7 Oct 04 04:16:19 crc kubenswrapper[4726]: I1004 04:16:19.300664 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" event={"ID":"e1087fbc-edc3-47d2-8229-fb1654ce31eb","Type":"ContainerStarted","Data":"561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7"} Oct 04 04:16:20 crc kubenswrapper[4726]: I1004 04:16:20.313699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" event={"ID":"e1087fbc-edc3-47d2-8229-fb1654ce31eb","Type":"ContainerStarted","Data":"e8341a8bd544b54292e2d6c6b3aed3a981ff5bbf5e9ada506934448751421c4f"} Oct 04 04:16:20 crc kubenswrapper[4726]: I1004 04:16:20.344156 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" podStartSLOduration=1.637872057 podStartE2EDuration="2.344136142s" podCreationTimestamp="2025-10-04 04:16:18 +0000 UTC" firstStartedPulling="2025-10-04 04:16:19.092719206 +0000 UTC m=+2153.267342429" lastFinishedPulling="2025-10-04 04:16:19.798983261 +0000 UTC m=+2153.973606514" observedRunningTime="2025-10-04 04:16:20.339972933 +0000 UTC m=+2154.514596186" watchObservedRunningTime="2025-10-04 04:16:20.344136142 +0000 UTC m=+2154.518759365" Oct 04 04:17:26 crc kubenswrapper[4726]: I1004 04:17:26.054632 4726 generic.go:334] "Generic (PLEG): container finished" podID="e1087fbc-edc3-47d2-8229-fb1654ce31eb" containerID="e8341a8bd544b54292e2d6c6b3aed3a981ff5bbf5e9ada506934448751421c4f" exitCode=0 Oct 04 04:17:26 crc kubenswrapper[4726]: I1004 04:17:26.054693 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" event={"ID":"e1087fbc-edc3-47d2-8229-fb1654ce31eb","Type":"ContainerDied","Data":"e8341a8bd544b54292e2d6c6b3aed3a981ff5bbf5e9ada506934448751421c4f"} Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.586124 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.657905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory\") pod \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.658152 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5v2l\" (UniqueName: \"kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l\") pod \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.658443 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key\") pod \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.658615 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0\") pod \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.658704 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle\") pod \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\" (UID: \"e1087fbc-edc3-47d2-8229-fb1654ce31eb\") " Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.670561 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e1087fbc-edc3-47d2-8229-fb1654ce31eb" (UID: "e1087fbc-edc3-47d2-8229-fb1654ce31eb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.670645 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l" (OuterVolumeSpecName: "kube-api-access-h5v2l") pod "e1087fbc-edc3-47d2-8229-fb1654ce31eb" (UID: "e1087fbc-edc3-47d2-8229-fb1654ce31eb"). InnerVolumeSpecName "kube-api-access-h5v2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.690891 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e1087fbc-edc3-47d2-8229-fb1654ce31eb" (UID: "e1087fbc-edc3-47d2-8229-fb1654ce31eb"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.700369 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory" (OuterVolumeSpecName: "inventory") pod "e1087fbc-edc3-47d2-8229-fb1654ce31eb" (UID: "e1087fbc-edc3-47d2-8229-fb1654ce31eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.701645 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e1087fbc-edc3-47d2-8229-fb1654ce31eb" (UID: "e1087fbc-edc3-47d2-8229-fb1654ce31eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.760942 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.760976 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5v2l\" (UniqueName: \"kubernetes.io/projected/e1087fbc-edc3-47d2-8229-fb1654ce31eb-kube-api-access-h5v2l\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.760989 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.760998 4726 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:27 crc kubenswrapper[4726]: I1004 04:17:27.761007 4726 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1087fbc-edc3-47d2-8229-fb1654ce31eb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.078334 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" event={"ID":"e1087fbc-edc3-47d2-8229-fb1654ce31eb","Type":"ContainerDied","Data":"561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7"} Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.078383 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="561b0e6bfc9889a445a05c08728ab7a6f711c57c87079ddb4d067fb67ac158c7" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.078399 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-jfxbp" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.195838 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz"] Oct 04 04:17:28 crc kubenswrapper[4726]: E1004 04:17:28.196310 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1087fbc-edc3-47d2-8229-fb1654ce31eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.196333 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1087fbc-edc3-47d2-8229-fb1654ce31eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.196575 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1087fbc-edc3-47d2-8229-fb1654ce31eb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.197316 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.199916 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.203033 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.203758 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.204244 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.204681 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.205042 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.214545 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz"] Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272206 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272361 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272398 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272466 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6nn4\" (UniqueName: \"kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.272537 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374415 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374516 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6nn4\" (UniqueName: \"kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374624 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374783 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.374836 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.380935 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.381629 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.383099 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.385052 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.388717 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.396401 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6nn4\" (UniqueName: \"kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:28 crc kubenswrapper[4726]: I1004 04:17:28.567710 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:17:29 crc kubenswrapper[4726]: I1004 04:17:29.196634 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz"] Oct 04 04:17:30 crc kubenswrapper[4726]: I1004 04:17:30.116055 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" event={"ID":"cbf08126-f813-40fe-b451-71cab2c2fd8a","Type":"ContainerStarted","Data":"6c7a21dea07ae09ee830b7cab9dac767168c0174c9e3ac5cf63c1f600ec5fdbd"} Oct 04 04:17:31 crc kubenswrapper[4726]: I1004 04:17:31.130333 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" event={"ID":"cbf08126-f813-40fe-b451-71cab2c2fd8a","Type":"ContainerStarted","Data":"87bddd2aa68b25c19dd2da2c73d0387a2032c7c4e37bfb2cc2c64f061255fbc0"} Oct 04 04:17:31 crc kubenswrapper[4726]: I1004 04:17:31.156853 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" podStartSLOduration=2.364619511 podStartE2EDuration="3.156824612s" podCreationTimestamp="2025-10-04 04:17:28 +0000 UTC" firstStartedPulling="2025-10-04 04:17:29.206391681 +0000 UTC m=+2223.381014934" lastFinishedPulling="2025-10-04 04:17:29.998596782 +0000 UTC m=+2224.173220035" observedRunningTime="2025-10-04 04:17:31.155435176 +0000 UTC m=+2225.330058489" watchObservedRunningTime="2025-10-04 04:17:31.156824612 +0000 UTC m=+2225.331447865" Oct 04 04:17:34 crc kubenswrapper[4726]: I1004 04:17:34.188550 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:17:34 crc kubenswrapper[4726]: I1004 04:17:34.189451 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:18:04 crc kubenswrapper[4726]: I1004 04:18:04.187942 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:18:04 crc kubenswrapper[4726]: I1004 04:18:04.188574 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:18:28 crc kubenswrapper[4726]: I1004 04:18:28.811087 4726 generic.go:334] "Generic (PLEG): container finished" podID="cbf08126-f813-40fe-b451-71cab2c2fd8a" containerID="87bddd2aa68b25c19dd2da2c73d0387a2032c7c4e37bfb2cc2c64f061255fbc0" exitCode=0 Oct 04 04:18:28 crc kubenswrapper[4726]: I1004 04:18:28.811276 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" event={"ID":"cbf08126-f813-40fe-b451-71cab2c2fd8a","Type":"ContainerDied","Data":"87bddd2aa68b25c19dd2da2c73d0387a2032c7c4e37bfb2cc2c64f061255fbc0"} Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.342649 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453379 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6nn4\" (UniqueName: \"kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453517 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453639 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453664 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453714 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.453751 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle\") pod \"cbf08126-f813-40fe-b451-71cab2c2fd8a\" (UID: \"cbf08126-f813-40fe-b451-71cab2c2fd8a\") " Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.460771 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4" (OuterVolumeSpecName: "kube-api-access-q6nn4") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "kube-api-access-q6nn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.461280 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.490538 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.502856 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory" (OuterVolumeSpecName: "inventory") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.509267 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.514373 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbf08126-f813-40fe-b451-71cab2c2fd8a" (UID: "cbf08126-f813-40fe-b451-71cab2c2fd8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555826 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555880 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555899 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555918 4726 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555940 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6nn4\" (UniqueName: \"kubernetes.io/projected/cbf08126-f813-40fe-b451-71cab2c2fd8a-kube-api-access-q6nn4\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.555959 4726 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cbf08126-f813-40fe-b451-71cab2c2fd8a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.838049 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" event={"ID":"cbf08126-f813-40fe-b451-71cab2c2fd8a","Type":"ContainerDied","Data":"6c7a21dea07ae09ee830b7cab9dac767168c0174c9e3ac5cf63c1f600ec5fdbd"} Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.838181 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c7a21dea07ae09ee830b7cab9dac767168c0174c9e3ac5cf63c1f600ec5fdbd" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.838222 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.964211 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d"] Oct 04 04:18:30 crc kubenswrapper[4726]: E1004 04:18:30.964832 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf08126-f813-40fe-b451-71cab2c2fd8a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.964858 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf08126-f813-40fe-b451-71cab2c2fd8a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.965202 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf08126-f813-40fe-b451-71cab2c2fd8a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.966216 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.968746 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.969639 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.969900 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.970186 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.973045 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:18:30 crc kubenswrapper[4726]: I1004 04:18:30.973314 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d"] Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.067825 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.068287 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmfw\" (UniqueName: \"kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.068471 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.068700 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.068931 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.170437 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.170748 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.170927 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmfw\" (UniqueName: \"kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.171404 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.171538 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.177960 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.179322 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.180659 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.188259 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.194996 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmfw\" (UniqueName: \"kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.317252 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:18:31 crc kubenswrapper[4726]: I1004 04:18:31.944771 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d"] Oct 04 04:18:32 crc kubenswrapper[4726]: I1004 04:18:32.858701 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" event={"ID":"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107","Type":"ContainerStarted","Data":"331a9fb4c75acd504ffcd41f5157a1139eaae50f66e952ced623654d4a9d8658"} Oct 04 04:18:32 crc kubenswrapper[4726]: I1004 04:18:32.859054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" event={"ID":"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107","Type":"ContainerStarted","Data":"9ac5882222e6a0fab9cbf3e0fd0b07ca493e8bb6de784f82b6fec68e624e035e"} Oct 04 04:18:32 crc kubenswrapper[4726]: I1004 04:18:32.915240 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" podStartSLOduration=2.399651755 podStartE2EDuration="2.915213096s" podCreationTimestamp="2025-10-04 04:18:30 +0000 UTC" firstStartedPulling="2025-10-04 04:18:31.971003054 +0000 UTC m=+2286.145626287" lastFinishedPulling="2025-10-04 04:18:32.486564415 +0000 UTC m=+2286.661187628" observedRunningTime="2025-10-04 04:18:32.911566925 +0000 UTC m=+2287.086190138" watchObservedRunningTime="2025-10-04 04:18:32.915213096 +0000 UTC m=+2287.089836349" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.188391 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.190023 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.190285 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.191427 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.191688 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" gracePeriod=600 Oct 04 04:18:34 crc kubenswrapper[4726]: E1004 04:18:34.319172 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.885785 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" exitCode=0 Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.885832 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7"} Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.885863 4726 scope.go:117] "RemoveContainer" containerID="a31a26626e3d18d938bd3fb7a8370b8e3ae55d40ce46684bcd3a5879257a98cb" Oct 04 04:18:34 crc kubenswrapper[4726]: I1004 04:18:34.886913 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:18:34 crc kubenswrapper[4726]: E1004 04:18:34.887640 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:18:46 crc kubenswrapper[4726]: I1004 04:18:46.521234 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:18:46 crc kubenswrapper[4726]: E1004 04:18:46.522022 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:19:01 crc kubenswrapper[4726]: I1004 04:19:01.502752 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:19:01 crc kubenswrapper[4726]: E1004 04:19:01.503473 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:19:14 crc kubenswrapper[4726]: I1004 04:19:14.504056 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:19:14 crc kubenswrapper[4726]: E1004 04:19:14.505396 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:19:26 crc kubenswrapper[4726]: I1004 04:19:26.508603 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:19:26 crc kubenswrapper[4726]: E1004 04:19:26.509356 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:19:41 crc kubenswrapper[4726]: I1004 04:19:41.503392 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:19:41 crc kubenswrapper[4726]: E1004 04:19:41.504488 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:19:53 crc kubenswrapper[4726]: I1004 04:19:53.503208 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:19:53 crc kubenswrapper[4726]: E1004 04:19:53.504481 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:20:05 crc kubenswrapper[4726]: I1004 04:20:05.504897 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:20:05 crc kubenswrapper[4726]: E1004 04:20:05.505861 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:20:16 crc kubenswrapper[4726]: I1004 04:20:16.515003 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:20:16 crc kubenswrapper[4726]: E1004 04:20:16.515980 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:20:31 crc kubenswrapper[4726]: I1004 04:20:31.503303 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:20:31 crc kubenswrapper[4726]: E1004 04:20:31.504321 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:20:44 crc kubenswrapper[4726]: I1004 04:20:44.514068 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:20:44 crc kubenswrapper[4726]: E1004 04:20:44.524214 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:20:59 crc kubenswrapper[4726]: I1004 04:20:59.503681 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:20:59 crc kubenswrapper[4726]: E1004 04:20:59.504929 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:21:12 crc kubenswrapper[4726]: I1004 04:21:12.502713 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:21:12 crc kubenswrapper[4726]: E1004 04:21:12.503660 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:21:24 crc kubenswrapper[4726]: I1004 04:21:24.502910 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:21:24 crc kubenswrapper[4726]: E1004 04:21:24.503771 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:21:37 crc kubenswrapper[4726]: I1004 04:21:37.502938 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:21:37 crc kubenswrapper[4726]: E1004 04:21:37.504116 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:21:48 crc kubenswrapper[4726]: I1004 04:21:48.503519 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:21:48 crc kubenswrapper[4726]: E1004 04:21:48.504951 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:22:01 crc kubenswrapper[4726]: I1004 04:22:01.503070 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:22:01 crc kubenswrapper[4726]: E1004 04:22:01.504168 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:22:13 crc kubenswrapper[4726]: I1004 04:22:13.503253 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:22:13 crc kubenswrapper[4726]: E1004 04:22:13.504042 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:22:28 crc kubenswrapper[4726]: I1004 04:22:28.502652 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:22:28 crc kubenswrapper[4726]: E1004 04:22:28.503573 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:22:41 crc kubenswrapper[4726]: I1004 04:22:41.502746 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:22:41 crc kubenswrapper[4726]: E1004 04:22:41.503704 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:22:56 crc kubenswrapper[4726]: I1004 04:22:56.523968 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:22:56 crc kubenswrapper[4726]: E1004 04:22:56.525642 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:23:07 crc kubenswrapper[4726]: I1004 04:23:07.503140 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:23:07 crc kubenswrapper[4726]: E1004 04:23:07.504240 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:23:15 crc kubenswrapper[4726]: I1004 04:23:15.173578 4726 generic.go:334] "Generic (PLEG): container finished" podID="836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" containerID="331a9fb4c75acd504ffcd41f5157a1139eaae50f66e952ced623654d4a9d8658" exitCode=0 Oct 04 04:23:15 crc kubenswrapper[4726]: I1004 04:23:15.173704 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" event={"ID":"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107","Type":"ContainerDied","Data":"331a9fb4c75acd504ffcd41f5157a1139eaae50f66e952ced623654d4a9d8658"} Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.749703 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.790495 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnmfw\" (UniqueName: \"kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw\") pod \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.790580 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle\") pod \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.790623 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory\") pod \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.790690 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key\") pod \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.790792 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0\") pod \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\" (UID: \"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107\") " Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.797544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw" (OuterVolumeSpecName: "kube-api-access-gnmfw") pod "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" (UID: "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107"). InnerVolumeSpecName "kube-api-access-gnmfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.803323 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" (UID: "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.826287 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" (UID: "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.830317 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory" (OuterVolumeSpecName: "inventory") pod "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" (UID: "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.841576 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" (UID: "836b7c28-6c4b-44c1-9f6f-ff52d2d7b107"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.892446 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnmfw\" (UniqueName: \"kubernetes.io/projected/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-kube-api-access-gnmfw\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.892481 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.892492 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.892501 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:16 crc kubenswrapper[4726]: I1004 04:23:16.892509 4726 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/836b7c28-6c4b-44c1-9f6f-ff52d2d7b107-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.201227 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" event={"ID":"836b7c28-6c4b-44c1-9f6f-ff52d2d7b107","Type":"ContainerDied","Data":"9ac5882222e6a0fab9cbf3e0fd0b07ca493e8bb6de784f82b6fec68e624e035e"} Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.201288 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac5882222e6a0fab9cbf3e0fd0b07ca493e8bb6de784f82b6fec68e624e035e" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.201347 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.338056 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8"] Oct 04 04:23:17 crc kubenswrapper[4726]: E1004 04:23:17.339308 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.339386 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.340613 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="836b7c28-6c4b-44c1-9f6f-ff52d2d7b107" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.342078 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.350185 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.350213 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.350862 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.351516 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.351978 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.352256 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.354242 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.379984 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8"] Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.400698 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.400794 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.400849 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.400902 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b54t2\" (UniqueName: \"kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.401002 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.401051 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.401210 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.401257 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.401308 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503400 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503469 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503566 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503596 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503676 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503701 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503737 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.503784 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b54t2\" (UniqueName: \"kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.505153 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.508412 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.508518 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.509194 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.509917 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.510119 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.510476 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.511137 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.525703 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b54t2\" (UniqueName: \"kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-j6dh8\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:17 crc kubenswrapper[4726]: I1004 04:23:17.707602 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:23:18 crc kubenswrapper[4726]: I1004 04:23:18.274543 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8"] Oct 04 04:23:18 crc kubenswrapper[4726]: W1004 04:23:18.279924 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb83849c_9bbd_479b_9a49_e99ea5247ab1.slice/crio-5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24 WatchSource:0}: Error finding container 5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24: Status 404 returned error can't find the container with id 5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24 Oct 04 04:23:18 crc kubenswrapper[4726]: I1004 04:23:18.283251 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:23:19 crc kubenswrapper[4726]: I1004 04:23:19.227337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" event={"ID":"cb83849c-9bbd-479b-9a49-e99ea5247ab1","Type":"ContainerStarted","Data":"3dae10e4ab5563632cf1d0170bc0b65fa9642652552421962d2df9b605b55f05"} Oct 04 04:23:19 crc kubenswrapper[4726]: I1004 04:23:19.227716 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" event={"ID":"cb83849c-9bbd-479b-9a49-e99ea5247ab1","Type":"ContainerStarted","Data":"5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24"} Oct 04 04:23:19 crc kubenswrapper[4726]: I1004 04:23:19.257808 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" podStartSLOduration=1.8322855869999999 podStartE2EDuration="2.257786271s" podCreationTimestamp="2025-10-04 04:23:17 +0000 UTC" firstStartedPulling="2025-10-04 04:23:18.282832349 +0000 UTC m=+2572.457455572" lastFinishedPulling="2025-10-04 04:23:18.708333003 +0000 UTC m=+2572.882956256" observedRunningTime="2025-10-04 04:23:19.254731864 +0000 UTC m=+2573.429355097" watchObservedRunningTime="2025-10-04 04:23:19.257786271 +0000 UTC m=+2573.432409484" Oct 04 04:23:20 crc kubenswrapper[4726]: I1004 04:23:20.506247 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:23:20 crc kubenswrapper[4726]: E1004 04:23:20.506887 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:23:34 crc kubenswrapper[4726]: I1004 04:23:34.504148 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:23:35 crc kubenswrapper[4726]: I1004 04:23:35.424467 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b"} Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.292543 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.295303 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.317604 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.344589 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.344739 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8kw8\" (UniqueName: \"kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.344830 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.447067 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.447242 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.447340 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8kw8\" (UniqueName: \"kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.447723 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.447831 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.476035 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8kw8\" (UniqueName: \"kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8\") pod \"community-operators-frsth\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:12 crc kubenswrapper[4726]: I1004 04:25:12.624495 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:13 crc kubenswrapper[4726]: I1004 04:25:13.193611 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:13 crc kubenswrapper[4726]: I1004 04:25:13.527819 4726 generic.go:334] "Generic (PLEG): container finished" podID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerID="8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af" exitCode=0 Oct 04 04:25:13 crc kubenswrapper[4726]: I1004 04:25:13.527879 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerDied","Data":"8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af"} Oct 04 04:25:13 crc kubenswrapper[4726]: I1004 04:25:13.528076 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerStarted","Data":"bcac4d16f48dbc5c42e77ed3776b7c8666156a6d964e972c80de61c9757f7176"} Oct 04 04:25:15 crc kubenswrapper[4726]: I1004 04:25:15.554809 4726 generic.go:334] "Generic (PLEG): container finished" podID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerID="46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3" exitCode=0 Oct 04 04:25:15 crc kubenswrapper[4726]: I1004 04:25:15.554890 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerDied","Data":"46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3"} Oct 04 04:25:16 crc kubenswrapper[4726]: I1004 04:25:16.571228 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerStarted","Data":"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8"} Oct 04 04:25:16 crc kubenswrapper[4726]: I1004 04:25:16.601264 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-frsth" podStartSLOduration=2.053412218 podStartE2EDuration="4.601246733s" podCreationTimestamp="2025-10-04 04:25:12 +0000 UTC" firstStartedPulling="2025-10-04 04:25:13.530990495 +0000 UTC m=+2687.705613708" lastFinishedPulling="2025-10-04 04:25:16.07882499 +0000 UTC m=+2690.253448223" observedRunningTime="2025-10-04 04:25:16.596352217 +0000 UTC m=+2690.770975460" watchObservedRunningTime="2025-10-04 04:25:16.601246733 +0000 UTC m=+2690.775869956" Oct 04 04:25:22 crc kubenswrapper[4726]: I1004 04:25:22.624967 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:22 crc kubenswrapper[4726]: I1004 04:25:22.625661 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:22 crc kubenswrapper[4726]: I1004 04:25:22.696715 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:22 crc kubenswrapper[4726]: I1004 04:25:22.780050 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:22 crc kubenswrapper[4726]: I1004 04:25:22.954231 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:24 crc kubenswrapper[4726]: I1004 04:25:24.677964 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-frsth" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="registry-server" containerID="cri-o://586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8" gracePeriod=2 Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.231162 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.339533 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content\") pod \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.339644 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8kw8\" (UniqueName: \"kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8\") pod \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.339709 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities\") pod \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\" (UID: \"2a44e34b-689a-4b3a-85ac-be57ff66fb2b\") " Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.340492 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities" (OuterVolumeSpecName: "utilities") pod "2a44e34b-689a-4b3a-85ac-be57ff66fb2b" (UID: "2a44e34b-689a-4b3a-85ac-be57ff66fb2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.346140 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8" (OuterVolumeSpecName: "kube-api-access-v8kw8") pod "2a44e34b-689a-4b3a-85ac-be57ff66fb2b" (UID: "2a44e34b-689a-4b3a-85ac-be57ff66fb2b"). InnerVolumeSpecName "kube-api-access-v8kw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.441605 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8kw8\" (UniqueName: \"kubernetes.io/projected/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-kube-api-access-v8kw8\") on node \"crc\" DevicePath \"\"" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.441636 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.554238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a44e34b-689a-4b3a-85ac-be57ff66fb2b" (UID: "2a44e34b-689a-4b3a-85ac-be57ff66fb2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.645775 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a44e34b-689a-4b3a-85ac-be57ff66fb2b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.687966 4726 generic.go:334] "Generic (PLEG): container finished" podID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerID="586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8" exitCode=0 Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.688012 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerDied","Data":"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8"} Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.688027 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-frsth" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.688045 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-frsth" event={"ID":"2a44e34b-689a-4b3a-85ac-be57ff66fb2b","Type":"ContainerDied","Data":"bcac4d16f48dbc5c42e77ed3776b7c8666156a6d964e972c80de61c9757f7176"} Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.688065 4726 scope.go:117] "RemoveContainer" containerID="586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.709717 4726 scope.go:117] "RemoveContainer" containerID="46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.727203 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.734905 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-frsth"] Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.745479 4726 scope.go:117] "RemoveContainer" containerID="8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.784261 4726 scope.go:117] "RemoveContainer" containerID="586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8" Oct 04 04:25:25 crc kubenswrapper[4726]: E1004 04:25:25.784724 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8\": container with ID starting with 586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8 not found: ID does not exist" containerID="586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.784757 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8"} err="failed to get container status \"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8\": rpc error: code = NotFound desc = could not find container \"586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8\": container with ID starting with 586593af547215dc5ea5506f0c26fcf84017287573b5d0ea5e32a6f244bfbda8 not found: ID does not exist" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.784785 4726 scope.go:117] "RemoveContainer" containerID="46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3" Oct 04 04:25:25 crc kubenswrapper[4726]: E1004 04:25:25.785066 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3\": container with ID starting with 46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3 not found: ID does not exist" containerID="46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.785182 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3"} err="failed to get container status \"46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3\": rpc error: code = NotFound desc = could not find container \"46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3\": container with ID starting with 46b88ad1fb5a9963f5ff9d0df2569805e7050f219c40715601b369188c4b14d3 not found: ID does not exist" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.785251 4726 scope.go:117] "RemoveContainer" containerID="8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af" Oct 04 04:25:25 crc kubenswrapper[4726]: E1004 04:25:25.785622 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af\": container with ID starting with 8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af not found: ID does not exist" containerID="8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af" Oct 04 04:25:25 crc kubenswrapper[4726]: I1004 04:25:25.785690 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af"} err="failed to get container status \"8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af\": rpc error: code = NotFound desc = could not find container \"8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af\": container with ID starting with 8a10d7763abccb517a0bc992e21661c76d1d6ba0559f2b0531932e4a93e391af not found: ID does not exist" Oct 04 04:25:26 crc kubenswrapper[4726]: I1004 04:25:26.543206 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" path="/var/lib/kubelet/pods/2a44e34b-689a-4b3a-85ac-be57ff66fb2b/volumes" Oct 04 04:25:34 crc kubenswrapper[4726]: I1004 04:25:34.189069 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:25:34 crc kubenswrapper[4726]: I1004 04:25:34.191271 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:26:04 crc kubenswrapper[4726]: I1004 04:26:04.189401 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:26:04 crc kubenswrapper[4726]: I1004 04:26:04.190040 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.928556 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:05 crc kubenswrapper[4726]: E1004 04:26:05.929672 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="registry-server" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.929702 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="registry-server" Oct 04 04:26:05 crc kubenswrapper[4726]: E1004 04:26:05.929746 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="extract-content" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.929759 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="extract-content" Oct 04 04:26:05 crc kubenswrapper[4726]: E1004 04:26:05.929790 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="extract-utilities" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.929804 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="extract-utilities" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.930188 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a44e34b-689a-4b3a-85ac-be57ff66fb2b" containerName="registry-server" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.932719 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:05 crc kubenswrapper[4726]: I1004 04:26:05.948789 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.068903 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.069186 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl2vp\" (UniqueName: \"kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.069389 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.171149 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl2vp\" (UniqueName: \"kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.171223 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.171312 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.172096 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.172122 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.198922 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl2vp\" (UniqueName: \"kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp\") pod \"certified-operators-wgkmf\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.258263 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:06 crc kubenswrapper[4726]: I1004 04:26:06.782352 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:07 crc kubenswrapper[4726]: I1004 04:26:07.137373 4726 generic.go:334] "Generic (PLEG): container finished" podID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerID="6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684" exitCode=0 Oct 04 04:26:07 crc kubenswrapper[4726]: I1004 04:26:07.137421 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerDied","Data":"6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684"} Oct 04 04:26:07 crc kubenswrapper[4726]: I1004 04:26:07.137463 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerStarted","Data":"eaa7a617a3f988b2b2452073660bf8b1f50b2deb1e291d2c868a6581224767b6"} Oct 04 04:26:08 crc kubenswrapper[4726]: I1004 04:26:08.148518 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerStarted","Data":"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec"} Oct 04 04:26:09 crc kubenswrapper[4726]: I1004 04:26:09.166375 4726 generic.go:334] "Generic (PLEG): container finished" podID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerID="b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec" exitCode=0 Oct 04 04:26:09 crc kubenswrapper[4726]: I1004 04:26:09.166518 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerDied","Data":"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec"} Oct 04 04:26:10 crc kubenswrapper[4726]: I1004 04:26:10.178952 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerStarted","Data":"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e"} Oct 04 04:26:10 crc kubenswrapper[4726]: I1004 04:26:10.221185 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wgkmf" podStartSLOduration=2.71346185 podStartE2EDuration="5.221164391s" podCreationTimestamp="2025-10-04 04:26:05 +0000 UTC" firstStartedPulling="2025-10-04 04:26:07.139449308 +0000 UTC m=+2741.314072521" lastFinishedPulling="2025-10-04 04:26:09.647151849 +0000 UTC m=+2743.821775062" observedRunningTime="2025-10-04 04:26:10.209907511 +0000 UTC m=+2744.384530764" watchObservedRunningTime="2025-10-04 04:26:10.221164391 +0000 UTC m=+2744.395787644" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.314693 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.316941 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.331562 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.435777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.435886 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-552nv\" (UniqueName: \"kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.436175 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.537697 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-552nv\" (UniqueName: \"kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.537758 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.537867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.538304 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.538585 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.565259 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-552nv\" (UniqueName: \"kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv\") pod \"redhat-operators-rvmzz\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:12 crc kubenswrapper[4726]: I1004 04:26:12.643745 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:13 crc kubenswrapper[4726]: I1004 04:26:13.165068 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:13 crc kubenswrapper[4726]: W1004 04:26:13.170573 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f1e38aa_8958_4ddc_a8d1_267751087a14.slice/crio-14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db WatchSource:0}: Error finding container 14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db: Status 404 returned error can't find the container with id 14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db Oct 04 04:26:13 crc kubenswrapper[4726]: I1004 04:26:13.212384 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerStarted","Data":"14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db"} Oct 04 04:26:14 crc kubenswrapper[4726]: I1004 04:26:14.226467 4726 generic.go:334] "Generic (PLEG): container finished" podID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerID="69f2bed4573ffabf50c69389f63918b6e7b1e4ac3edf0ac1d0e6b259388a31df" exitCode=0 Oct 04 04:26:14 crc kubenswrapper[4726]: I1004 04:26:14.226717 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerDied","Data":"69f2bed4573ffabf50c69389f63918b6e7b1e4ac3edf0ac1d0e6b259388a31df"} Oct 04 04:26:15 crc kubenswrapper[4726]: I1004 04:26:15.239348 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerStarted","Data":"05603a9bd3ee23422849247ebbf9196564b4b54d1f225505194ea845aca017fa"} Oct 04 04:26:16 crc kubenswrapper[4726]: I1004 04:26:16.254077 4726 generic.go:334] "Generic (PLEG): container finished" podID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerID="05603a9bd3ee23422849247ebbf9196564b4b54d1f225505194ea845aca017fa" exitCode=0 Oct 04 04:26:16 crc kubenswrapper[4726]: I1004 04:26:16.254153 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerDied","Data":"05603a9bd3ee23422849247ebbf9196564b4b54d1f225505194ea845aca017fa"} Oct 04 04:26:16 crc kubenswrapper[4726]: I1004 04:26:16.260026 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:16 crc kubenswrapper[4726]: I1004 04:26:16.260223 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:16 crc kubenswrapper[4726]: I1004 04:26:16.338072 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:17 crc kubenswrapper[4726]: I1004 04:26:17.269269 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerStarted","Data":"1b4306f5f2c643e5edf0dfdc64e999d642d358e7d063be4dcac60de621f13c75"} Oct 04 04:26:17 crc kubenswrapper[4726]: I1004 04:26:17.310883 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rvmzz" podStartSLOduration=2.64653769 podStartE2EDuration="5.310848057s" podCreationTimestamp="2025-10-04 04:26:12 +0000 UTC" firstStartedPulling="2025-10-04 04:26:14.232647654 +0000 UTC m=+2748.407270907" lastFinishedPulling="2025-10-04 04:26:16.896958031 +0000 UTC m=+2751.071581274" observedRunningTime="2025-10-04 04:26:17.295318787 +0000 UTC m=+2751.469942090" watchObservedRunningTime="2025-10-04 04:26:17.310848057 +0000 UTC m=+2751.485471320" Oct 04 04:26:17 crc kubenswrapper[4726]: I1004 04:26:17.349303 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:18 crc kubenswrapper[4726]: I1004 04:26:18.698365 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.296274 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wgkmf" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="registry-server" containerID="cri-o://ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e" gracePeriod=2 Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.525436 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.528068 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.559904 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.627166 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhxgn\" (UniqueName: \"kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.627221 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.627260 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.730455 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhxgn\" (UniqueName: \"kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.730520 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.730556 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.731123 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.731385 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.752530 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhxgn\" (UniqueName: \"kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn\") pod \"redhat-marketplace-tbnqh\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.838042 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:20 crc kubenswrapper[4726]: I1004 04:26:20.865679 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.037725 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl2vp\" (UniqueName: \"kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp\") pod \"81f23399-238e-4ad8-9ac1-0a16de45dabe\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.038125 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content\") pod \"81f23399-238e-4ad8-9ac1-0a16de45dabe\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.038167 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities\") pod \"81f23399-238e-4ad8-9ac1-0a16de45dabe\" (UID: \"81f23399-238e-4ad8-9ac1-0a16de45dabe\") " Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.038882 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities" (OuterVolumeSpecName: "utilities") pod "81f23399-238e-4ad8-9ac1-0a16de45dabe" (UID: "81f23399-238e-4ad8-9ac1-0a16de45dabe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.044324 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp" (OuterVolumeSpecName: "kube-api-access-wl2vp") pod "81f23399-238e-4ad8-9ac1-0a16de45dabe" (UID: "81f23399-238e-4ad8-9ac1-0a16de45dabe"). InnerVolumeSpecName "kube-api-access-wl2vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.083552 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81f23399-238e-4ad8-9ac1-0a16de45dabe" (UID: "81f23399-238e-4ad8-9ac1-0a16de45dabe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.140528 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl2vp\" (UniqueName: \"kubernetes.io/projected/81f23399-238e-4ad8-9ac1-0a16de45dabe-kube-api-access-wl2vp\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.140556 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.140566 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f23399-238e-4ad8-9ac1-0a16de45dabe-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.310273 4726 generic.go:334] "Generic (PLEG): container finished" podID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerID="ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e" exitCode=0 Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.310332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerDied","Data":"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e"} Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.310365 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgkmf" event={"ID":"81f23399-238e-4ad8-9ac1-0a16de45dabe","Type":"ContainerDied","Data":"eaa7a617a3f988b2b2452073660bf8b1f50b2deb1e291d2c868a6581224767b6"} Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.310390 4726 scope.go:117] "RemoveContainer" containerID="ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.310565 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgkmf" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.337646 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.363734 4726 scope.go:117] "RemoveContainer" containerID="b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.364795 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.382386 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wgkmf"] Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.391128 4726 scope.go:117] "RemoveContainer" containerID="6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.426505 4726 scope.go:117] "RemoveContainer" containerID="ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e" Oct 04 04:26:21 crc kubenswrapper[4726]: E1004 04:26:21.426990 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e\": container with ID starting with ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e not found: ID does not exist" containerID="ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.427039 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e"} err="failed to get container status \"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e\": rpc error: code = NotFound desc = could not find container \"ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e\": container with ID starting with ff49de2b6a3b9abac254dc147fb6eb86b78394a137e29b49a176b6a9db0c420e not found: ID does not exist" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.427067 4726 scope.go:117] "RemoveContainer" containerID="b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec" Oct 04 04:26:21 crc kubenswrapper[4726]: E1004 04:26:21.427890 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec\": container with ID starting with b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec not found: ID does not exist" containerID="b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.427931 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec"} err="failed to get container status \"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec\": rpc error: code = NotFound desc = could not find container \"b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec\": container with ID starting with b55e7bc1fdaf9c849317c30416b402bce528b2ad9ad4a16dd86c26624b526cec not found: ID does not exist" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.427959 4726 scope.go:117] "RemoveContainer" containerID="6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684" Oct 04 04:26:21 crc kubenswrapper[4726]: E1004 04:26:21.428328 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684\": container with ID starting with 6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684 not found: ID does not exist" containerID="6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684" Oct 04 04:26:21 crc kubenswrapper[4726]: I1004 04:26:21.428375 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684"} err="failed to get container status \"6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684\": rpc error: code = NotFound desc = could not find container \"6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684\": container with ID starting with 6ebfa887f4b742169c3ebfc1efc2ae4e01e3744540d6a39e441f66171acc0684 not found: ID does not exist" Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.319515 4726 generic.go:334] "Generic (PLEG): container finished" podID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerID="8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc" exitCode=0 Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.319652 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerDied","Data":"8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc"} Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.319895 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerStarted","Data":"548b909f995d230c18ae086de087719383cba6606bfc9e31c0239ffabdd3cd5a"} Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.518476 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" path="/var/lib/kubelet/pods/81f23399-238e-4ad8-9ac1-0a16de45dabe/volumes" Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.644602 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.644932 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:22 crc kubenswrapper[4726]: I1004 04:26:22.726477 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:23 crc kubenswrapper[4726]: I1004 04:26:23.347287 4726 generic.go:334] "Generic (PLEG): container finished" podID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerID="f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc" exitCode=0 Oct 04 04:26:23 crc kubenswrapper[4726]: I1004 04:26:23.347431 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerDied","Data":"f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc"} Oct 04 04:26:23 crc kubenswrapper[4726]: I1004 04:26:23.412350 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:24 crc kubenswrapper[4726]: I1004 04:26:24.364260 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerStarted","Data":"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb"} Oct 04 04:26:24 crc kubenswrapper[4726]: I1004 04:26:24.392863 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tbnqh" podStartSLOduration=2.923064931 podStartE2EDuration="4.392844845s" podCreationTimestamp="2025-10-04 04:26:20 +0000 UTC" firstStartedPulling="2025-10-04 04:26:22.321696944 +0000 UTC m=+2756.496320157" lastFinishedPulling="2025-10-04 04:26:23.791476828 +0000 UTC m=+2757.966100071" observedRunningTime="2025-10-04 04:26:24.389606772 +0000 UTC m=+2758.564229985" watchObservedRunningTime="2025-10-04 04:26:24.392844845 +0000 UTC m=+2758.567468078" Oct 04 04:26:26 crc kubenswrapper[4726]: I1004 04:26:26.897400 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:26 crc kubenswrapper[4726]: I1004 04:26:26.897893 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rvmzz" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="registry-server" containerID="cri-o://1b4306f5f2c643e5edf0dfdc64e999d642d358e7d063be4dcac60de621f13c75" gracePeriod=2 Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.400214 4726 generic.go:334] "Generic (PLEG): container finished" podID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerID="1b4306f5f2c643e5edf0dfdc64e999d642d358e7d063be4dcac60de621f13c75" exitCode=0 Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.400638 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerDied","Data":"1b4306f5f2c643e5edf0dfdc64e999d642d358e7d063be4dcac60de621f13c75"} Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.400688 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rvmzz" event={"ID":"5f1e38aa-8958-4ddc-a8d1-267751087a14","Type":"ContainerDied","Data":"14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db"} Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.400708 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14b165ccddd28a6f955df1f8716b525a20cfc7a2dff023505470460514ef63db" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.460020 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.575915 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-552nv\" (UniqueName: \"kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv\") pod \"5f1e38aa-8958-4ddc-a8d1-267751087a14\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.576154 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content\") pod \"5f1e38aa-8958-4ddc-a8d1-267751087a14\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.576289 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities\") pod \"5f1e38aa-8958-4ddc-a8d1-267751087a14\" (UID: \"5f1e38aa-8958-4ddc-a8d1-267751087a14\") " Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.578555 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities" (OuterVolumeSpecName: "utilities") pod "5f1e38aa-8958-4ddc-a8d1-267751087a14" (UID: "5f1e38aa-8958-4ddc-a8d1-267751087a14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.586588 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv" (OuterVolumeSpecName: "kube-api-access-552nv") pod "5f1e38aa-8958-4ddc-a8d1-267751087a14" (UID: "5f1e38aa-8958-4ddc-a8d1-267751087a14"). InnerVolumeSpecName "kube-api-access-552nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.651854 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f1e38aa-8958-4ddc-a8d1-267751087a14" (UID: "5f1e38aa-8958-4ddc-a8d1-267751087a14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.678661 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.678687 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f1e38aa-8958-4ddc-a8d1-267751087a14-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:27 crc kubenswrapper[4726]: I1004 04:26:27.678696 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-552nv\" (UniqueName: \"kubernetes.io/projected/5f1e38aa-8958-4ddc-a8d1-267751087a14-kube-api-access-552nv\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:28 crc kubenswrapper[4726]: I1004 04:26:28.414989 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rvmzz" Oct 04 04:26:28 crc kubenswrapper[4726]: I1004 04:26:28.459865 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:28 crc kubenswrapper[4726]: I1004 04:26:28.466769 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rvmzz"] Oct 04 04:26:28 crc kubenswrapper[4726]: I1004 04:26:28.511959 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" path="/var/lib/kubelet/pods/5f1e38aa-8958-4ddc-a8d1-267751087a14/volumes" Oct 04 04:26:30 crc kubenswrapper[4726]: I1004 04:26:30.867888 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:30 crc kubenswrapper[4726]: I1004 04:26:30.868834 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:30 crc kubenswrapper[4726]: I1004 04:26:30.960307 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:31 crc kubenswrapper[4726]: I1004 04:26:31.521088 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:32 crc kubenswrapper[4726]: I1004 04:26:32.533847 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:33 crc kubenswrapper[4726]: I1004 04:26:33.473799 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tbnqh" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="registry-server" containerID="cri-o://e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb" gracePeriod=2 Oct 04 04:26:33 crc kubenswrapper[4726]: I1004 04:26:33.944068 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.028527 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities\") pod \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.028636 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhxgn\" (UniqueName: \"kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn\") pod \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.028767 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content\") pod \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\" (UID: \"b950d33c-aba0-4a18-a228-5d0e1ff01b44\") " Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.030439 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities" (OuterVolumeSpecName: "utilities") pod "b950d33c-aba0-4a18-a228-5d0e1ff01b44" (UID: "b950d33c-aba0-4a18-a228-5d0e1ff01b44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.035678 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn" (OuterVolumeSpecName: "kube-api-access-jhxgn") pod "b950d33c-aba0-4a18-a228-5d0e1ff01b44" (UID: "b950d33c-aba0-4a18-a228-5d0e1ff01b44"). InnerVolumeSpecName "kube-api-access-jhxgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.046733 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b950d33c-aba0-4a18-a228-5d0e1ff01b44" (UID: "b950d33c-aba0-4a18-a228-5d0e1ff01b44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.131283 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.131320 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhxgn\" (UniqueName: \"kubernetes.io/projected/b950d33c-aba0-4a18-a228-5d0e1ff01b44-kube-api-access-jhxgn\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.131334 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b950d33c-aba0-4a18-a228-5d0e1ff01b44-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.188507 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.188597 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.188658 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.189559 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.189676 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b" gracePeriod=600 Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.506269 4726 generic.go:334] "Generic (PLEG): container finished" podID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerID="e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb" exitCode=0 Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.506784 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbnqh" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.511663 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b" exitCode=0 Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.522054 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerDied","Data":"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb"} Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.522100 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbnqh" event={"ID":"b950d33c-aba0-4a18-a228-5d0e1ff01b44","Type":"ContainerDied","Data":"548b909f995d230c18ae086de087719383cba6606bfc9e31c0239ffabdd3cd5a"} Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.522159 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b"} Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.522184 4726 scope.go:117] "RemoveContainer" containerID="e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.548010 4726 scope.go:117] "RemoveContainer" containerID="f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.575252 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.580048 4726 scope.go:117] "RemoveContainer" containerID="8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.583469 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbnqh"] Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.649704 4726 scope.go:117] "RemoveContainer" containerID="e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb" Oct 04 04:26:34 crc kubenswrapper[4726]: E1004 04:26:34.650260 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb\": container with ID starting with e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb not found: ID does not exist" containerID="e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.650311 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb"} err="failed to get container status \"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb\": rpc error: code = NotFound desc = could not find container \"e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb\": container with ID starting with e1e4e15dfc97cca74286e1484527588482dc6bb0cbb2faae8c8d9f33c7b877fb not found: ID does not exist" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.650338 4726 scope.go:117] "RemoveContainer" containerID="f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc" Oct 04 04:26:34 crc kubenswrapper[4726]: E1004 04:26:34.650807 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc\": container with ID starting with f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc not found: ID does not exist" containerID="f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.650851 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc"} err="failed to get container status \"f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc\": rpc error: code = NotFound desc = could not find container \"f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc\": container with ID starting with f388c26e14946bd8774d3482bd0a947576dd6bb9aabcced42f181b0a1ee01dfc not found: ID does not exist" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.650869 4726 scope.go:117] "RemoveContainer" containerID="8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc" Oct 04 04:26:34 crc kubenswrapper[4726]: E1004 04:26:34.651251 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc\": container with ID starting with 8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc not found: ID does not exist" containerID="8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.651288 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc"} err="failed to get container status \"8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc\": rpc error: code = NotFound desc = could not find container \"8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc\": container with ID starting with 8e16f2e34e7136d3a8d2cb1110b53d3d3896f9742fbc05103f57531084fab3dc not found: ID does not exist" Oct 04 04:26:34 crc kubenswrapper[4726]: I1004 04:26:34.651305 4726 scope.go:117] "RemoveContainer" containerID="b0dec8da0c4ff46f548bc73fa180f19fede2c1588e50e0ec07e3de908ae5e8a7" Oct 04 04:26:35 crc kubenswrapper[4726]: I1004 04:26:35.536821 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815"} Oct 04 04:26:36 crc kubenswrapper[4726]: I1004 04:26:36.518555 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" path="/var/lib/kubelet/pods/b950d33c-aba0-4a18-a228-5d0e1ff01b44/volumes" Oct 04 04:26:40 crc kubenswrapper[4726]: I1004 04:26:40.613157 4726 generic.go:334] "Generic (PLEG): container finished" podID="cb83849c-9bbd-479b-9a49-e99ea5247ab1" containerID="3dae10e4ab5563632cf1d0170bc0b65fa9642652552421962d2df9b605b55f05" exitCode=0 Oct 04 04:26:40 crc kubenswrapper[4726]: I1004 04:26:40.613213 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" event={"ID":"cb83849c-9bbd-479b-9a49-e99ea5247ab1","Type":"ContainerDied","Data":"3dae10e4ab5563632cf1d0170bc0b65fa9642652552421962d2df9b605b55f05"} Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.044783 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.099030 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.099949 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100145 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100169 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100186 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100301 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100336 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b54t2\" (UniqueName: \"kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100361 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.100384 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle\") pod \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\" (UID: \"cb83849c-9bbd-479b-9a49-e99ea5247ab1\") " Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.111943 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2" (OuterVolumeSpecName: "kube-api-access-b54t2") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "kube-api-access-b54t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.134543 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.147540 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.149821 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.150431 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.151137 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.156182 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.161530 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.168604 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory" (OuterVolumeSpecName: "inventory") pod "cb83849c-9bbd-479b-9a49-e99ea5247ab1" (UID: "cb83849c-9bbd-479b-9a49-e99ea5247ab1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203360 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203395 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203426 4726 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203436 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203447 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b54t2\" (UniqueName: \"kubernetes.io/projected/cb83849c-9bbd-479b-9a49-e99ea5247ab1-kube-api-access-b54t2\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203456 4726 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203464 4726 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203472 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.203479 4726 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cb83849c-9bbd-479b-9a49-e99ea5247ab1-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.637739 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" event={"ID":"cb83849c-9bbd-479b-9a49-e99ea5247ab1","Type":"ContainerDied","Data":"5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24"} Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.637775 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f6a63f97002ae4c7ad2170d9b74574995fea1e1c1499c9af783161d10e75e24" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.637845 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-j6dh8" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.757864 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9"] Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758330 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758349 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758363 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758372 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758388 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758398 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758423 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758431 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758446 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb83849c-9bbd-479b-9a49-e99ea5247ab1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758454 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb83849c-9bbd-479b-9a49-e99ea5247ab1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758471 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758479 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758493 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758501 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758517 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758527 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="extract-content" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758546 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758554 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: E1004 04:26:42.758567 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758575 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="extract-utilities" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758824 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b950d33c-aba0-4a18-a228-5d0e1ff01b44" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758838 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f23399-238e-4ad8-9ac1-0a16de45dabe" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758858 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f1e38aa-8958-4ddc-a8d1-267751087a14" containerName="registry-server" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.758872 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb83849c-9bbd-479b-9a49-e99ea5247ab1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.759639 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.761968 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.762577 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-xzj84" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.762821 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.763094 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.764470 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.781501 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9"] Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816586 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816633 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816652 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816673 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816710 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816734 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.816767 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g56md\" (UniqueName: \"kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918326 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918390 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918410 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918433 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918486 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918511 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.918542 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g56md\" (UniqueName: \"kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.923035 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.923066 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.923266 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.923709 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.923894 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.925795 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:42 crc kubenswrapper[4726]: I1004 04:26:42.936220 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g56md\" (UniqueName: \"kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wclx9\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:43 crc kubenswrapper[4726]: I1004 04:26:43.080120 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:26:43 crc kubenswrapper[4726]: I1004 04:26:43.451086 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9"] Oct 04 04:26:43 crc kubenswrapper[4726]: I1004 04:26:43.651867 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" event={"ID":"e9a93781-fcc5-4323-bb02-934800e8de8f","Type":"ContainerStarted","Data":"ead63eb3a12b9e447cd4ca61718a122ab1b61e46446c650f6ae75a51d2181029"} Oct 04 04:26:44 crc kubenswrapper[4726]: I1004 04:26:44.664071 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" event={"ID":"e9a93781-fcc5-4323-bb02-934800e8de8f","Type":"ContainerStarted","Data":"cf0df2e58f5a5d664152cd4f612ade91f308d28d95094bca6cd0e5c91f591861"} Oct 04 04:26:44 crc kubenswrapper[4726]: I1004 04:26:44.689817 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" podStartSLOduration=2.169184371 podStartE2EDuration="2.689791466s" podCreationTimestamp="2025-10-04 04:26:42 +0000 UTC" firstStartedPulling="2025-10-04 04:26:43.446001194 +0000 UTC m=+2777.620624407" lastFinishedPulling="2025-10-04 04:26:43.966608289 +0000 UTC m=+2778.141231502" observedRunningTime="2025-10-04 04:26:44.683698529 +0000 UTC m=+2778.858321752" watchObservedRunningTime="2025-10-04 04:26:44.689791466 +0000 UTC m=+2778.864414679" Oct 04 04:28:34 crc kubenswrapper[4726]: I1004 04:28:34.188032 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:28:34 crc kubenswrapper[4726]: I1004 04:28:34.188619 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:29:04 crc kubenswrapper[4726]: I1004 04:29:04.188991 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:29:04 crc kubenswrapper[4726]: I1004 04:29:04.189574 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:29:27 crc kubenswrapper[4726]: I1004 04:29:27.291876 4726 generic.go:334] "Generic (PLEG): container finished" podID="e9a93781-fcc5-4323-bb02-934800e8de8f" containerID="cf0df2e58f5a5d664152cd4f612ade91f308d28d95094bca6cd0e5c91f591861" exitCode=0 Oct 04 04:29:27 crc kubenswrapper[4726]: I1004 04:29:27.291970 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" event={"ID":"e9a93781-fcc5-4323-bb02-934800e8de8f","Type":"ContainerDied","Data":"cf0df2e58f5a5d664152cd4f612ade91f308d28d95094bca6cd0e5c91f591861"} Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.731544 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.841636 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.841784 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.841836 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.841898 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.841970 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g56md\" (UniqueName: \"kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.842074 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.842176 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory\") pod \"e9a93781-fcc5-4323-bb02-934800e8de8f\" (UID: \"e9a93781-fcc5-4323-bb02-934800e8de8f\") " Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.847990 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md" (OuterVolumeSpecName: "kube-api-access-g56md") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "kube-api-access-g56md". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.860936 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.871349 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.872855 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory" (OuterVolumeSpecName: "inventory") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.877046 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.879551 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.880365 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e9a93781-fcc5-4323-bb02-934800e8de8f" (UID: "e9a93781-fcc5-4323-bb02-934800e8de8f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.944833 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945091 4726 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945271 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945408 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945550 4726 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945738 4726 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e9a93781-fcc5-4323-bb02-934800e8de8f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:28 crc kubenswrapper[4726]: I1004 04:29:28.945914 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g56md\" (UniqueName: \"kubernetes.io/projected/e9a93781-fcc5-4323-bb02-934800e8de8f-kube-api-access-g56md\") on node \"crc\" DevicePath \"\"" Oct 04 04:29:29 crc kubenswrapper[4726]: I1004 04:29:29.311355 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" event={"ID":"e9a93781-fcc5-4323-bb02-934800e8de8f","Type":"ContainerDied","Data":"ead63eb3a12b9e447cd4ca61718a122ab1b61e46446c650f6ae75a51d2181029"} Oct 04 04:29:29 crc kubenswrapper[4726]: I1004 04:29:29.311712 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ead63eb3a12b9e447cd4ca61718a122ab1b61e46446c650f6ae75a51d2181029" Oct 04 04:29:29 crc kubenswrapper[4726]: I1004 04:29:29.311431 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wclx9" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.188461 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.188930 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.188985 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.189846 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.189916 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" gracePeriod=600 Oct 04 04:29:34 crc kubenswrapper[4726]: E1004 04:29:34.336940 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.357998 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" exitCode=0 Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.358046 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815"} Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.358148 4726 scope.go:117] "RemoveContainer" containerID="498941957af9b0c06cc8679e8bdf8b3b2b29b55aad846e18992b501ce6479c7b" Oct 04 04:29:34 crc kubenswrapper[4726]: I1004 04:29:34.358827 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:29:34 crc kubenswrapper[4726]: E1004 04:29:34.360717 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:29:47 crc kubenswrapper[4726]: I1004 04:29:47.503674 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:29:47 crc kubenswrapper[4726]: E1004 04:29:47.505336 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.181681 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4"] Oct 04 04:30:00 crc kubenswrapper[4726]: E1004 04:30:00.182711 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a93781-fcc5-4323-bb02-934800e8de8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.182726 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a93781-fcc5-4323-bb02-934800e8de8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.182914 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a93781-fcc5-4323-bb02-934800e8de8f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.183593 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.187739 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.188140 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.196010 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4"] Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.310366 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.310435 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxmqz\" (UniqueName: \"kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.310493 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.412408 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.412490 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxmqz\" (UniqueName: \"kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.412559 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.413909 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.419708 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.430382 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxmqz\" (UniqueName: \"kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz\") pod \"collect-profiles-29325870-zcpn4\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.520026 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:00 crc kubenswrapper[4726]: I1004 04:30:00.984811 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4"] Oct 04 04:30:01 crc kubenswrapper[4726]: I1004 04:30:01.651418 4726 generic.go:334] "Generic (PLEG): container finished" podID="937c5c22-b2a5-4aa3-af8e-faca307c4799" containerID="9794d592e5ade475a231d5c6310cdec72a118b11fd23d94c5b1c2a883ca42889" exitCode=0 Oct 04 04:30:01 crc kubenswrapper[4726]: I1004 04:30:01.651500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" event={"ID":"937c5c22-b2a5-4aa3-af8e-faca307c4799","Type":"ContainerDied","Data":"9794d592e5ade475a231d5c6310cdec72a118b11fd23d94c5b1c2a883ca42889"} Oct 04 04:30:01 crc kubenswrapper[4726]: I1004 04:30:01.651907 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" event={"ID":"937c5c22-b2a5-4aa3-af8e-faca307c4799","Type":"ContainerStarted","Data":"7c3b98510b00784348ad69bbbcb7699a2dc5218a4ee4c0fac0aea0d3d55b883a"} Oct 04 04:30:02 crc kubenswrapper[4726]: I1004 04:30:02.502537 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:30:02 crc kubenswrapper[4726]: E1004 04:30:02.503317 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.048038 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.210238 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume\") pod \"937c5c22-b2a5-4aa3-af8e-faca307c4799\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.210584 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxmqz\" (UniqueName: \"kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz\") pod \"937c5c22-b2a5-4aa3-af8e-faca307c4799\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.210803 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume\") pod \"937c5c22-b2a5-4aa3-af8e-faca307c4799\" (UID: \"937c5c22-b2a5-4aa3-af8e-faca307c4799\") " Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.211446 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume" (OuterVolumeSpecName: "config-volume") pod "937c5c22-b2a5-4aa3-af8e-faca307c4799" (UID: "937c5c22-b2a5-4aa3-af8e-faca307c4799"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.211610 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/937c5c22-b2a5-4aa3-af8e-faca307c4799-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.216646 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz" (OuterVolumeSpecName: "kube-api-access-zxmqz") pod "937c5c22-b2a5-4aa3-af8e-faca307c4799" (UID: "937c5c22-b2a5-4aa3-af8e-faca307c4799"). InnerVolumeSpecName "kube-api-access-zxmqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.217368 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "937c5c22-b2a5-4aa3-af8e-faca307c4799" (UID: "937c5c22-b2a5-4aa3-af8e-faca307c4799"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.313227 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxmqz\" (UniqueName: \"kubernetes.io/projected/937c5c22-b2a5-4aa3-af8e-faca307c4799-kube-api-access-zxmqz\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.313268 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/937c5c22-b2a5-4aa3-af8e-faca307c4799-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.673422 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" event={"ID":"937c5c22-b2a5-4aa3-af8e-faca307c4799","Type":"ContainerDied","Data":"7c3b98510b00784348ad69bbbcb7699a2dc5218a4ee4c0fac0aea0d3d55b883a"} Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.673462 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c3b98510b00784348ad69bbbcb7699a2dc5218a4ee4c0fac0aea0d3d55b883a" Oct 04 04:30:03 crc kubenswrapper[4726]: I1004 04:30:03.673489 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-zcpn4" Oct 04 04:30:03 crc kubenswrapper[4726]: E1004 04:30:03.813949 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod937c5c22_b2a5_4aa3_af8e_faca307c4799.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod937c5c22_b2a5_4aa3_af8e_faca307c4799.slice/crio-7c3b98510b00784348ad69bbbcb7699a2dc5218a4ee4c0fac0aea0d3d55b883a\": RecentStats: unable to find data in memory cache]" Oct 04 04:30:04 crc kubenswrapper[4726]: I1004 04:30:04.130957 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn"] Oct 04 04:30:04 crc kubenswrapper[4726]: I1004 04:30:04.140364 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325825-df8cn"] Oct 04 04:30:04 crc kubenswrapper[4726]: I1004 04:30:04.536631 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf" path="/var/lib/kubelet/pods/1db8acf4-4e94-4518-a8b3-0a4beb0aa1bf/volumes" Oct 04 04:30:16 crc kubenswrapper[4726]: I1004 04:30:16.512377 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:30:16 crc kubenswrapper[4726]: E1004 04:30:16.513270 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:30:20 crc kubenswrapper[4726]: I1004 04:30:20.561171 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5bdb856b6c-f5xfb" podUID="503e5303-a481-4184-9bb0-2369123d5267" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.427829 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 04:30:26 crc kubenswrapper[4726]: E1004 04:30:26.428932 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="937c5c22-b2a5-4aa3-af8e-faca307c4799" containerName="collect-profiles" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.428945 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="937c5c22-b2a5-4aa3-af8e-faca307c4799" containerName="collect-profiles" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.429194 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="937c5c22-b2a5-4aa3-af8e-faca307c4799" containerName="collect-profiles" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.430089 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.432593 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.432701 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.432721 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nkvlq" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.433122 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.438371 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610491 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610536 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610569 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610659 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2jhs\" (UniqueName: \"kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610700 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610724 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610741 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.610910 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.713870 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.713930 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.713954 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.713977 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714000 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714058 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714077 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714118 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714179 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2jhs\" (UniqueName: \"kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.714551 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.715087 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.716002 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.716084 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.717146 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.717704 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.722141 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.723219 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.726173 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.729559 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.732077 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2jhs\" (UniqueName: \"kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.742774 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"tempest-tests-tempest\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " pod="openstack/tempest-tests-tempest" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.763564 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nkvlq" Oct 04 04:30:26 crc kubenswrapper[4726]: I1004 04:30:26.771571 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 04:30:27 crc kubenswrapper[4726]: I1004 04:30:27.075938 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 04:30:27 crc kubenswrapper[4726]: I1004 04:30:27.081385 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:30:27 crc kubenswrapper[4726]: I1004 04:30:27.957857 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de0aaf35-392c-4705-b554-1d4768c5cff1","Type":"ContainerStarted","Data":"e8b53caaec78e26a806c8f0e218f57da8a24e6feed7d6ba14458b6a89036cd07"} Oct 04 04:30:29 crc kubenswrapper[4726]: I1004 04:30:29.503330 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:30:29 crc kubenswrapper[4726]: E1004 04:30:29.503916 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:30:38 crc kubenswrapper[4726]: I1004 04:30:38.073867 4726 scope.go:117] "RemoveContainer" containerID="bd005a81442dc0147dc57f59565cf94f9d07985fa1244cccd45fac5e3e332a23" Oct 04 04:30:41 crc kubenswrapper[4726]: I1004 04:30:41.503377 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:30:41 crc kubenswrapper[4726]: E1004 04:30:41.504665 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:30:53 crc kubenswrapper[4726]: E1004 04:30:53.000817 4726 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 04 04:30:53 crc kubenswrapper[4726]: E1004 04:30:53.001405 4726 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n2jhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(de0aaf35-392c-4705-b554-1d4768c5cff1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:30:53 crc kubenswrapper[4726]: E1004 04:30:53.002712 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="de0aaf35-392c-4705-b554-1d4768c5cff1" Oct 04 04:30:53 crc kubenswrapper[4726]: E1004 04:30:53.209509 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="de0aaf35-392c-4705-b554-1d4768c5cff1" Oct 04 04:30:54 crc kubenswrapper[4726]: I1004 04:30:54.503367 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:30:54 crc kubenswrapper[4726]: E1004 04:30:54.504208 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:31:04 crc kubenswrapper[4726]: I1004 04:31:04.275542 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 04:31:06 crc kubenswrapper[4726]: I1004 04:31:06.344702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de0aaf35-392c-4705-b554-1d4768c5cff1","Type":"ContainerStarted","Data":"16770c96268e50fcfc974cbbf6862eee0beccb42481a2fee2f994fa7cadd29f0"} Oct 04 04:31:06 crc kubenswrapper[4726]: I1004 04:31:06.368752 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.176727841 podStartE2EDuration="41.368736559s" podCreationTimestamp="2025-10-04 04:30:25 +0000 UTC" firstStartedPulling="2025-10-04 04:30:27.080995881 +0000 UTC m=+3001.255619114" lastFinishedPulling="2025-10-04 04:31:04.273004619 +0000 UTC m=+3038.447627832" observedRunningTime="2025-10-04 04:31:06.368168024 +0000 UTC m=+3040.542791277" watchObservedRunningTime="2025-10-04 04:31:06.368736559 +0000 UTC m=+3040.543359772" Oct 04 04:31:08 crc kubenswrapper[4726]: I1004 04:31:08.503476 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:31:08 crc kubenswrapper[4726]: E1004 04:31:08.504165 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:31:23 crc kubenswrapper[4726]: I1004 04:31:23.503284 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:31:23 crc kubenswrapper[4726]: E1004 04:31:23.504602 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:31:36 crc kubenswrapper[4726]: I1004 04:31:36.509395 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:31:36 crc kubenswrapper[4726]: E1004 04:31:36.510181 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:31:47 crc kubenswrapper[4726]: I1004 04:31:47.502067 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:31:47 crc kubenswrapper[4726]: E1004 04:31:47.502846 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:31:58 crc kubenswrapper[4726]: I1004 04:31:58.503218 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:31:58 crc kubenswrapper[4726]: E1004 04:31:58.503979 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:32:09 crc kubenswrapper[4726]: I1004 04:32:09.502984 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:32:09 crc kubenswrapper[4726]: E1004 04:32:09.503811 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:32:21 crc kubenswrapper[4726]: I1004 04:32:21.503037 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:32:21 crc kubenswrapper[4726]: E1004 04:32:21.504042 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:32:36 crc kubenswrapper[4726]: I1004 04:32:36.513696 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:32:36 crc kubenswrapper[4726]: E1004 04:32:36.514770 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:32:49 crc kubenswrapper[4726]: I1004 04:32:49.503648 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:32:49 crc kubenswrapper[4726]: E1004 04:32:49.504509 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:32:52 crc kubenswrapper[4726]: I1004 04:32:52.957850 4726 scope.go:117] "RemoveContainer" containerID="69f2bed4573ffabf50c69389f63918b6e7b1e4ac3edf0ac1d0e6b259388a31df" Oct 04 04:32:52 crc kubenswrapper[4726]: I1004 04:32:52.987370 4726 scope.go:117] "RemoveContainer" containerID="05603a9bd3ee23422849247ebbf9196564b4b54d1f225505194ea845aca017fa" Oct 04 04:32:53 crc kubenswrapper[4726]: I1004 04:32:53.055311 4726 scope.go:117] "RemoveContainer" containerID="1b4306f5f2c643e5edf0dfdc64e999d642d358e7d063be4dcac60de621f13c75" Oct 04 04:33:01 crc kubenswrapper[4726]: I1004 04:33:01.503063 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:33:01 crc kubenswrapper[4726]: E1004 04:33:01.504098 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:33:12 crc kubenswrapper[4726]: I1004 04:33:12.502468 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:33:12 crc kubenswrapper[4726]: E1004 04:33:12.503308 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:33:26 crc kubenswrapper[4726]: I1004 04:33:26.510010 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:33:26 crc kubenswrapper[4726]: E1004 04:33:26.510931 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:33:41 crc kubenswrapper[4726]: I1004 04:33:41.502988 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:33:41 crc kubenswrapper[4726]: E1004 04:33:41.504249 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:33:56 crc kubenswrapper[4726]: I1004 04:33:56.509349 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:33:56 crc kubenswrapper[4726]: E1004 04:33:56.510203 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:34:07 crc kubenswrapper[4726]: I1004 04:34:07.502877 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:34:07 crc kubenswrapper[4726]: E1004 04:34:07.504766 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:34:22 crc kubenswrapper[4726]: I1004 04:34:22.502914 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:34:22 crc kubenswrapper[4726]: E1004 04:34:22.504369 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:34:37 crc kubenswrapper[4726]: I1004 04:34:37.502971 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:34:38 crc kubenswrapper[4726]: I1004 04:34:38.364500 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f"} Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.340130 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.342708 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.382564 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.542497 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.542881 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2kt9\" (UniqueName: \"kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.543348 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.645331 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.645452 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.645487 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2kt9\" (UniqueName: \"kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.645841 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.646062 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.676150 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2kt9\" (UniqueName: \"kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9\") pod \"community-operators-858vr\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:53 crc kubenswrapper[4726]: I1004 04:35:53.695873 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:35:54 crc kubenswrapper[4726]: I1004 04:35:54.245742 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:35:55 crc kubenswrapper[4726]: I1004 04:35:55.076791 4726 generic.go:334] "Generic (PLEG): container finished" podID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerID="802592ee88f7ad8a4410e31fd8a2c398830b4275772fe558ef9cd83797223be6" exitCode=0 Oct 04 04:35:55 crc kubenswrapper[4726]: I1004 04:35:55.076903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerDied","Data":"802592ee88f7ad8a4410e31fd8a2c398830b4275772fe558ef9cd83797223be6"} Oct 04 04:35:55 crc kubenswrapper[4726]: I1004 04:35:55.077149 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerStarted","Data":"5b054146e03295d0396353371d71c556b140cc877de02a829fc4d1465a2d0e30"} Oct 04 04:35:55 crc kubenswrapper[4726]: I1004 04:35:55.080693 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:35:58 crc kubenswrapper[4726]: I1004 04:35:58.102254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerStarted","Data":"faeca389e987375e5f54005a099774764cd9ef4bb5ed802416d1130b9cc02d23"} Oct 04 04:35:59 crc kubenswrapper[4726]: I1004 04:35:59.115498 4726 generic.go:334] "Generic (PLEG): container finished" podID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerID="faeca389e987375e5f54005a099774764cd9ef4bb5ed802416d1130b9cc02d23" exitCode=0 Oct 04 04:35:59 crc kubenswrapper[4726]: I1004 04:35:59.115669 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerDied","Data":"faeca389e987375e5f54005a099774764cd9ef4bb5ed802416d1130b9cc02d23"} Oct 04 04:36:05 crc kubenswrapper[4726]: I1004 04:36:05.170229 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerStarted","Data":"58d92d5fb533a0825aeb44e1e33fbc16ff9cf28360c311561bf236d9bd455a55"} Oct 04 04:36:05 crc kubenswrapper[4726]: I1004 04:36:05.195570 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-858vr" podStartSLOduration=3.420835405 podStartE2EDuration="12.195542806s" podCreationTimestamp="2025-10-04 04:35:53 +0000 UTC" firstStartedPulling="2025-10-04 04:35:55.080450385 +0000 UTC m=+3329.255073598" lastFinishedPulling="2025-10-04 04:36:03.855157776 +0000 UTC m=+3338.029780999" observedRunningTime="2025-10-04 04:36:05.187529081 +0000 UTC m=+3339.362152294" watchObservedRunningTime="2025-10-04 04:36:05.195542806 +0000 UTC m=+3339.370166019" Oct 04 04:36:13 crc kubenswrapper[4726]: I1004 04:36:13.696425 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:13 crc kubenswrapper[4726]: I1004 04:36:13.697086 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:13 crc kubenswrapper[4726]: I1004 04:36:13.749796 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:14 crc kubenswrapper[4726]: I1004 04:36:14.301704 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:14 crc kubenswrapper[4726]: I1004 04:36:14.350418 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:36:16 crc kubenswrapper[4726]: I1004 04:36:16.274990 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-858vr" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="registry-server" containerID="cri-o://58d92d5fb533a0825aeb44e1e33fbc16ff9cf28360c311561bf236d9bd455a55" gracePeriod=2 Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.293064 4726 generic.go:334] "Generic (PLEG): container finished" podID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerID="58d92d5fb533a0825aeb44e1e33fbc16ff9cf28360c311561bf236d9bd455a55" exitCode=0 Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.293160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerDied","Data":"58d92d5fb533a0825aeb44e1e33fbc16ff9cf28360c311561bf236d9bd455a55"} Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.784495 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.944217 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2kt9\" (UniqueName: \"kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9\") pod \"41d62431-c35e-4a7a-860a-ece26409c1cc\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.944287 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities\") pod \"41d62431-c35e-4a7a-860a-ece26409c1cc\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.945252 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities" (OuterVolumeSpecName: "utilities") pod "41d62431-c35e-4a7a-860a-ece26409c1cc" (UID: "41d62431-c35e-4a7a-860a-ece26409c1cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.945924 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content\") pod \"41d62431-c35e-4a7a-860a-ece26409c1cc\" (UID: \"41d62431-c35e-4a7a-860a-ece26409c1cc\") " Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.946521 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:18 crc kubenswrapper[4726]: I1004 04:36:18.950328 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9" (OuterVolumeSpecName: "kube-api-access-b2kt9") pod "41d62431-c35e-4a7a-860a-ece26409c1cc" (UID: "41d62431-c35e-4a7a-860a-ece26409c1cc"). InnerVolumeSpecName "kube-api-access-b2kt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.008084 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41d62431-c35e-4a7a-860a-ece26409c1cc" (UID: "41d62431-c35e-4a7a-860a-ece26409c1cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.048704 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2kt9\" (UniqueName: \"kubernetes.io/projected/41d62431-c35e-4a7a-860a-ece26409c1cc-kube-api-access-b2kt9\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.048762 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d62431-c35e-4a7a-860a-ece26409c1cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.303618 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-858vr" event={"ID":"41d62431-c35e-4a7a-860a-ece26409c1cc","Type":"ContainerDied","Data":"5b054146e03295d0396353371d71c556b140cc877de02a829fc4d1465a2d0e30"} Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.303665 4726 scope.go:117] "RemoveContainer" containerID="58d92d5fb533a0825aeb44e1e33fbc16ff9cf28360c311561bf236d9bd455a55" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.303675 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-858vr" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.332677 4726 scope.go:117] "RemoveContainer" containerID="faeca389e987375e5f54005a099774764cd9ef4bb5ed802416d1130b9cc02d23" Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.342744 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.350994 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-858vr"] Oct 04 04:36:19 crc kubenswrapper[4726]: I1004 04:36:19.371374 4726 scope.go:117] "RemoveContainer" containerID="802592ee88f7ad8a4410e31fd8a2c398830b4275772fe558ef9cd83797223be6" Oct 04 04:36:20 crc kubenswrapper[4726]: I1004 04:36:20.521781 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" path="/var/lib/kubelet/pods/41d62431-c35e-4a7a-860a-ece26409c1cc/volumes" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.631844 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:36:38 crc kubenswrapper[4726]: E1004 04:36:38.633555 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="extract-content" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.633579 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="extract-content" Oct 04 04:36:38 crc kubenswrapper[4726]: E1004 04:36:38.633592 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="registry-server" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.633603 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="registry-server" Oct 04 04:36:38 crc kubenswrapper[4726]: E1004 04:36:38.633666 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="extract-utilities" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.633678 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="extract-utilities" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.633985 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d62431-c35e-4a7a-860a-ece26409c1cc" containerName="registry-server" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.636834 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.640922 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.729918 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9vhj\" (UniqueName: \"kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.730025 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.730631 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.832695 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.832766 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9vhj\" (UniqueName: \"kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.832867 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.833384 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.833512 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.853445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9vhj\" (UniqueName: \"kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj\") pod \"redhat-operators-xc5s7\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:38 crc kubenswrapper[4726]: I1004 04:36:38.969091 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:39 crc kubenswrapper[4726]: I1004 04:36:39.429341 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:36:39 crc kubenswrapper[4726]: W1004 04:36:39.449821 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f421c7b_0857_4a94_a358_2a67b183ba9d.slice/crio-49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e WatchSource:0}: Error finding container 49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e: Status 404 returned error can't find the container with id 49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e Oct 04 04:36:39 crc kubenswrapper[4726]: I1004 04:36:39.484078 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerStarted","Data":"49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e"} Oct 04 04:36:40 crc kubenswrapper[4726]: I1004 04:36:40.493978 4726 generic.go:334] "Generic (PLEG): container finished" podID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerID="33eb50e92869fe10a60b422e60f3ea9777c2e89394fcb239b0cdaa747aabc805" exitCode=0 Oct 04 04:36:40 crc kubenswrapper[4726]: I1004 04:36:40.494058 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerDied","Data":"33eb50e92869fe10a60b422e60f3ea9777c2e89394fcb239b0cdaa747aabc805"} Oct 04 04:36:44 crc kubenswrapper[4726]: I1004 04:36:44.549903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerStarted","Data":"0e814e7a35be3d228208aa9cd093798c5cdfbdf5fc355ff0a7b09470293310d0"} Oct 04 04:36:45 crc kubenswrapper[4726]: I1004 04:36:45.596870 4726 generic.go:334] "Generic (PLEG): container finished" podID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerID="0e814e7a35be3d228208aa9cd093798c5cdfbdf5fc355ff0a7b09470293310d0" exitCode=0 Oct 04 04:36:45 crc kubenswrapper[4726]: I1004 04:36:45.597200 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerDied","Data":"0e814e7a35be3d228208aa9cd093798c5cdfbdf5fc355ff0a7b09470293310d0"} Oct 04 04:36:53 crc kubenswrapper[4726]: I1004 04:36:53.691725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerStarted","Data":"661bb426475da1a1fb0f6d7468c9a18dc25c301565609099cf65af8dde98820a"} Oct 04 04:36:53 crc kubenswrapper[4726]: I1004 04:36:53.715581 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xc5s7" podStartSLOduration=3.098654696 podStartE2EDuration="15.715563357s" podCreationTimestamp="2025-10-04 04:36:38 +0000 UTC" firstStartedPulling="2025-10-04 04:36:40.496641839 +0000 UTC m=+3374.671265062" lastFinishedPulling="2025-10-04 04:36:53.11355051 +0000 UTC m=+3387.288173723" observedRunningTime="2025-10-04 04:36:53.712090788 +0000 UTC m=+3387.886714001" watchObservedRunningTime="2025-10-04 04:36:53.715563357 +0000 UTC m=+3387.890186570" Oct 04 04:36:58 crc kubenswrapper[4726]: I1004 04:36:58.969333 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:58 crc kubenswrapper[4726]: I1004 04:36:58.970141 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:59 crc kubenswrapper[4726]: I1004 04:36:59.038084 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:59 crc kubenswrapper[4726]: I1004 04:36:59.813010 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:36:59 crc kubenswrapper[4726]: I1004 04:36:59.868649 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:37:01 crc kubenswrapper[4726]: I1004 04:37:01.778290 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xc5s7" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="registry-server" containerID="cri-o://661bb426475da1a1fb0f6d7468c9a18dc25c301565609099cf65af8dde98820a" gracePeriod=2 Oct 04 04:37:02 crc kubenswrapper[4726]: I1004 04:37:02.793563 4726 generic.go:334] "Generic (PLEG): container finished" podID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerID="661bb426475da1a1fb0f6d7468c9a18dc25c301565609099cf65af8dde98820a" exitCode=0 Oct 04 04:37:02 crc kubenswrapper[4726]: I1004 04:37:02.793632 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerDied","Data":"661bb426475da1a1fb0f6d7468c9a18dc25c301565609099cf65af8dde98820a"} Oct 04 04:37:02 crc kubenswrapper[4726]: I1004 04:37:02.793661 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xc5s7" event={"ID":"3f421c7b-0857-4a94-a358-2a67b183ba9d","Type":"ContainerDied","Data":"49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e"} Oct 04 04:37:02 crc kubenswrapper[4726]: I1004 04:37:02.793673 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49d1f677fcbc47118d78cc3d3e9984c6ebfd8de28c4fc10a9b1a60dc9fca137e" Oct 04 04:37:02 crc kubenswrapper[4726]: I1004 04:37:02.881087 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.011853 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities\") pod \"3f421c7b-0857-4a94-a358-2a67b183ba9d\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.012281 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content\") pod \"3f421c7b-0857-4a94-a358-2a67b183ba9d\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.012319 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9vhj\" (UniqueName: \"kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj\") pod \"3f421c7b-0857-4a94-a358-2a67b183ba9d\" (UID: \"3f421c7b-0857-4a94-a358-2a67b183ba9d\") " Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.013444 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities" (OuterVolumeSpecName: "utilities") pod "3f421c7b-0857-4a94-a358-2a67b183ba9d" (UID: "3f421c7b-0857-4a94-a358-2a67b183ba9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.023519 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj" (OuterVolumeSpecName: "kube-api-access-q9vhj") pod "3f421c7b-0857-4a94-a358-2a67b183ba9d" (UID: "3f421c7b-0857-4a94-a358-2a67b183ba9d"). InnerVolumeSpecName "kube-api-access-q9vhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.114807 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.114852 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9vhj\" (UniqueName: \"kubernetes.io/projected/3f421c7b-0857-4a94-a358-2a67b183ba9d-kube-api-access-q9vhj\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.118238 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f421c7b-0857-4a94-a358-2a67b183ba9d" (UID: "3f421c7b-0857-4a94-a358-2a67b183ba9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.216842 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f421c7b-0857-4a94-a358-2a67b183ba9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.802045 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xc5s7" Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.845754 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:37:03 crc kubenswrapper[4726]: I1004 04:37:03.861033 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xc5s7"] Oct 04 04:37:04 crc kubenswrapper[4726]: I1004 04:37:04.187774 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:37:04 crc kubenswrapper[4726]: I1004 04:37:04.187836 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:37:04 crc kubenswrapper[4726]: I1004 04:37:04.518587 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" path="/var/lib/kubelet/pods/3f421c7b-0857-4a94-a358-2a67b183ba9d/volumes" Oct 04 04:37:34 crc kubenswrapper[4726]: I1004 04:37:34.188036 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:37:34 crc kubenswrapper[4726]: I1004 04:37:34.190396 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.188853 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.189863 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.189960 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.191295 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.191412 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f" gracePeriod=600 Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.390475 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f" exitCode=0 Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.390557 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f"} Oct 04 04:38:04 crc kubenswrapper[4726]: I1004 04:38:04.390948 4726 scope.go:117] "RemoveContainer" containerID="c1674d9335febbe073ed04d75c6475231ce097fc7ca84139bcc6015f09af9815" Oct 04 04:38:05 crc kubenswrapper[4726]: I1004 04:38:05.405063 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911"} Oct 04 04:40:04 crc kubenswrapper[4726]: I1004 04:40:04.188177 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:40:04 crc kubenswrapper[4726]: I1004 04:40:04.188865 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:40:34 crc kubenswrapper[4726]: I1004 04:40:34.188151 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:40:34 crc kubenswrapper[4726]: I1004 04:40:34.188741 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.057425 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:40:54 crc kubenswrapper[4726]: E1004 04:40:54.058833 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="extract-utilities" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.058857 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="extract-utilities" Oct 04 04:40:54 crc kubenswrapper[4726]: E1004 04:40:54.058910 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="extract-content" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.058920 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="extract-content" Oct 04 04:40:54 crc kubenswrapper[4726]: E1004 04:40:54.058934 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="registry-server" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.058942 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="registry-server" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.059203 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f421c7b-0857-4a94-a358-2a67b183ba9d" containerName="registry-server" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.060826 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.066365 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.145785 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d8fw\" (UniqueName: \"kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.145947 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.145973 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.247648 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d8fw\" (UniqueName: \"kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.247747 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.247774 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.248312 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.248368 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.283060 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d8fw\" (UniqueName: \"kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw\") pod \"certified-operators-ldd2b\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.388309 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:40:54 crc kubenswrapper[4726]: I1004 04:40:54.923349 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:40:55 crc kubenswrapper[4726]: I1004 04:40:55.024166 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerStarted","Data":"708ace6d08d6b96bdaf8ff7563108f8b08a483609f27a84362a25cdb05dd0d1b"} Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.033959 4726 generic.go:334] "Generic (PLEG): container finished" podID="840b686c-580f-44aa-9d58-aff48accfbf1" containerID="a49d528227b4a5c6cddce3fbcebd358424ea6de34393f99c5d632728cee59072" exitCode=0 Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.034198 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerDied","Data":"a49d528227b4a5c6cddce3fbcebd358424ea6de34393f99c5d632728cee59072"} Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.036222 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.460806 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.463227 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.470614 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.597890 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.598187 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vch9w\" (UniqueName: \"kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.598675 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.700904 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.700976 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.700995 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vch9w\" (UniqueName: \"kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.701707 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.701786 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.727849 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vch9w\" (UniqueName: \"kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w\") pod \"redhat-marketplace-jdp8z\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:56 crc kubenswrapper[4726]: I1004 04:40:56.800031 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:40:57 crc kubenswrapper[4726]: I1004 04:40:57.412679 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:40:57 crc kubenswrapper[4726]: W1004 04:40:57.422344 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb492feec_89af_4e54_b0e6_012436301beb.slice/crio-b987d745ef209c1239acc00a78c6438c2110fc38b034405dc89799f6a732abbd WatchSource:0}: Error finding container b987d745ef209c1239acc00a78c6438c2110fc38b034405dc89799f6a732abbd: Status 404 returned error can't find the container with id b987d745ef209c1239acc00a78c6438c2110fc38b034405dc89799f6a732abbd Oct 04 04:40:58 crc kubenswrapper[4726]: I1004 04:40:58.050666 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerStarted","Data":"b987d745ef209c1239acc00a78c6438c2110fc38b034405dc89799f6a732abbd"} Oct 04 04:40:59 crc kubenswrapper[4726]: I1004 04:40:59.062834 4726 generic.go:334] "Generic (PLEG): container finished" podID="b492feec-89af-4e54-b0e6-012436301beb" containerID="254f72929504d9c66ba9891647c4dfb02a7905d03e015187a45e86f0c55f8b4b" exitCode=0 Oct 04 04:40:59 crc kubenswrapper[4726]: I1004 04:40:59.062993 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerDied","Data":"254f72929504d9c66ba9891647c4dfb02a7905d03e015187a45e86f0c55f8b4b"} Oct 04 04:41:00 crc kubenswrapper[4726]: I1004 04:41:00.074627 4726 generic.go:334] "Generic (PLEG): container finished" podID="840b686c-580f-44aa-9d58-aff48accfbf1" containerID="be9b8426dce24b697c44581c777f7f7b381c07a9fa2f9c6fa018371f35510d1f" exitCode=0 Oct 04 04:41:00 crc kubenswrapper[4726]: I1004 04:41:00.074692 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerDied","Data":"be9b8426dce24b697c44581c777f7f7b381c07a9fa2f9c6fa018371f35510d1f"} Oct 04 04:41:02 crc kubenswrapper[4726]: I1004 04:41:02.101126 4726 generic.go:334] "Generic (PLEG): container finished" podID="b492feec-89af-4e54-b0e6-012436301beb" containerID="41e6b8bc4bc5b6e07c17db318da30cd1b6f77f564c32fdbd368a79e609889a5d" exitCode=0 Oct 04 04:41:02 crc kubenswrapper[4726]: I1004 04:41:02.101216 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerDied","Data":"41e6b8bc4bc5b6e07c17db318da30cd1b6f77f564c32fdbd368a79e609889a5d"} Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.122324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerStarted","Data":"de8c610ffdbffaa7ae93238b92ddf3f7cb46e00397b81bcdb9853158cc562079"} Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.149229 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ldd2b" podStartSLOduration=3.463232288 podStartE2EDuration="10.149210247s" podCreationTimestamp="2025-10-04 04:40:54 +0000 UTC" firstStartedPulling="2025-10-04 04:40:56.035976355 +0000 UTC m=+3630.210599568" lastFinishedPulling="2025-10-04 04:41:02.721954314 +0000 UTC m=+3636.896577527" observedRunningTime="2025-10-04 04:41:04.142068795 +0000 UTC m=+3638.316692018" watchObservedRunningTime="2025-10-04 04:41:04.149210247 +0000 UTC m=+3638.323833470" Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.188697 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.188984 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.189200 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.192784 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.192980 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" gracePeriod=600 Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.389365 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:04 crc kubenswrapper[4726]: I1004 04:41:04.389786 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.132961 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" exitCode=0 Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.133032 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911"} Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.133065 4726 scope.go:117] "RemoveContainer" containerID="75eb06714c5455a97730988e3a0073726cf1fc3f25c9e020cfc6d6152fc2708f" Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.135836 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerStarted","Data":"95d7ce397fb8f06044fb1076eef334e8a08e20170facea128160b67a6e1f8477"} Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.154720 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jdp8z" podStartSLOduration=4.049885833 podStartE2EDuration="9.154697541s" podCreationTimestamp="2025-10-04 04:40:56 +0000 UTC" firstStartedPulling="2025-10-04 04:40:59.163669791 +0000 UTC m=+3633.338293004" lastFinishedPulling="2025-10-04 04:41:04.268481489 +0000 UTC m=+3638.443104712" observedRunningTime="2025-10-04 04:41:05.150006781 +0000 UTC m=+3639.324630014" watchObservedRunningTime="2025-10-04 04:41:05.154697541 +0000 UTC m=+3639.329320754" Oct 04 04:41:05 crc kubenswrapper[4726]: I1004 04:41:05.439351 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-ldd2b" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="registry-server" probeResult="failure" output=< Oct 04 04:41:05 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:41:05 crc kubenswrapper[4726]: > Oct 04 04:41:06 crc kubenswrapper[4726]: E1004 04:41:06.213902 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:41:06 crc kubenswrapper[4726]: I1004 04:41:06.800422 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:06 crc kubenswrapper[4726]: I1004 04:41:06.800827 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:06 crc kubenswrapper[4726]: I1004 04:41:06.850571 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:07 crc kubenswrapper[4726]: I1004 04:41:07.164862 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:41:07 crc kubenswrapper[4726]: E1004 04:41:07.165430 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:41:14 crc kubenswrapper[4726]: I1004 04:41:14.465693 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:14 crc kubenswrapper[4726]: I1004 04:41:14.517199 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:14 crc kubenswrapper[4726]: I1004 04:41:14.707420 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:41:16 crc kubenswrapper[4726]: I1004 04:41:16.254439 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ldd2b" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="registry-server" containerID="cri-o://de8c610ffdbffaa7ae93238b92ddf3f7cb46e00397b81bcdb9853158cc562079" gracePeriod=2 Oct 04 04:41:16 crc kubenswrapper[4726]: I1004 04:41:16.864123 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.107634 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.267618 4726 generic.go:334] "Generic (PLEG): container finished" podID="840b686c-580f-44aa-9d58-aff48accfbf1" containerID="de8c610ffdbffaa7ae93238b92ddf3f7cb46e00397b81bcdb9853158cc562079" exitCode=0 Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.267696 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerDied","Data":"de8c610ffdbffaa7ae93238b92ddf3f7cb46e00397b81bcdb9853158cc562079"} Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.267967 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jdp8z" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="registry-server" containerID="cri-o://95d7ce397fb8f06044fb1076eef334e8a08e20170facea128160b67a6e1f8477" gracePeriod=2 Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.839547 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.920218 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content\") pod \"840b686c-580f-44aa-9d58-aff48accfbf1\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.922365 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d8fw\" (UniqueName: \"kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw\") pod \"840b686c-580f-44aa-9d58-aff48accfbf1\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.922417 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities\") pod \"840b686c-580f-44aa-9d58-aff48accfbf1\" (UID: \"840b686c-580f-44aa-9d58-aff48accfbf1\") " Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.923022 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities" (OuterVolumeSpecName: "utilities") pod "840b686c-580f-44aa-9d58-aff48accfbf1" (UID: "840b686c-580f-44aa-9d58-aff48accfbf1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.923201 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:17 crc kubenswrapper[4726]: I1004 04:41:17.928781 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw" (OuterVolumeSpecName: "kube-api-access-7d8fw") pod "840b686c-580f-44aa-9d58-aff48accfbf1" (UID: "840b686c-580f-44aa-9d58-aff48accfbf1"). InnerVolumeSpecName "kube-api-access-7d8fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.025317 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d8fw\" (UniqueName: \"kubernetes.io/projected/840b686c-580f-44aa-9d58-aff48accfbf1-kube-api-access-7d8fw\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.240065 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "840b686c-580f-44aa-9d58-aff48accfbf1" (UID: "840b686c-580f-44aa-9d58-aff48accfbf1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.279616 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ldd2b" event={"ID":"840b686c-580f-44aa-9d58-aff48accfbf1","Type":"ContainerDied","Data":"708ace6d08d6b96bdaf8ff7563108f8b08a483609f27a84362a25cdb05dd0d1b"} Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.279655 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ldd2b" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.279710 4726 scope.go:117] "RemoveContainer" containerID="de8c610ffdbffaa7ae93238b92ddf3f7cb46e00397b81bcdb9853158cc562079" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.281750 4726 generic.go:334] "Generic (PLEG): container finished" podID="b492feec-89af-4e54-b0e6-012436301beb" containerID="95d7ce397fb8f06044fb1076eef334e8a08e20170facea128160b67a6e1f8477" exitCode=0 Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.281782 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerDied","Data":"95d7ce397fb8f06044fb1076eef334e8a08e20170facea128160b67a6e1f8477"} Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.315251 4726 scope.go:117] "RemoveContainer" containerID="be9b8426dce24b697c44581c777f7f7b381c07a9fa2f9c6fa018371f35510d1f" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.333272 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/840b686c-580f-44aa-9d58-aff48accfbf1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.343608 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.355792 4726 scope.go:117] "RemoveContainer" containerID="a49d528227b4a5c6cddce3fbcebd358424ea6de34393f99c5d632728cee59072" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.357361 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ldd2b"] Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.515312 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" path="/var/lib/kubelet/pods/840b686c-580f-44aa-9d58-aff48accfbf1/volumes" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.664704 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.843409 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities\") pod \"b492feec-89af-4e54-b0e6-012436301beb\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.843626 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content\") pod \"b492feec-89af-4e54-b0e6-012436301beb\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.843671 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vch9w\" (UniqueName: \"kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w\") pod \"b492feec-89af-4e54-b0e6-012436301beb\" (UID: \"b492feec-89af-4e54-b0e6-012436301beb\") " Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.844139 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities" (OuterVolumeSpecName: "utilities") pod "b492feec-89af-4e54-b0e6-012436301beb" (UID: "b492feec-89af-4e54-b0e6-012436301beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.844727 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.857622 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b492feec-89af-4e54-b0e6-012436301beb" (UID: "b492feec-89af-4e54-b0e6-012436301beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.860892 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w" (OuterVolumeSpecName: "kube-api-access-vch9w") pod "b492feec-89af-4e54-b0e6-012436301beb" (UID: "b492feec-89af-4e54-b0e6-012436301beb"). InnerVolumeSpecName "kube-api-access-vch9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.947754 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b492feec-89af-4e54-b0e6-012436301beb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:18 crc kubenswrapper[4726]: I1004 04:41:18.947790 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vch9w\" (UniqueName: \"kubernetes.io/projected/b492feec-89af-4e54-b0e6-012436301beb-kube-api-access-vch9w\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.293625 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jdp8z" event={"ID":"b492feec-89af-4e54-b0e6-012436301beb","Type":"ContainerDied","Data":"b987d745ef209c1239acc00a78c6438c2110fc38b034405dc89799f6a732abbd"} Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.293649 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jdp8z" Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.293944 4726 scope.go:117] "RemoveContainer" containerID="95d7ce397fb8f06044fb1076eef334e8a08e20170facea128160b67a6e1f8477" Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.312521 4726 scope.go:117] "RemoveContainer" containerID="41e6b8bc4bc5b6e07c17db318da30cd1b6f77f564c32fdbd368a79e609889a5d" Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.329188 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.335068 4726 scope.go:117] "RemoveContainer" containerID="254f72929504d9c66ba9891647c4dfb02a7905d03e015187a45e86f0c55f8b4b" Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.336283 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jdp8z"] Oct 04 04:41:19 crc kubenswrapper[4726]: I1004 04:41:19.503087 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:41:19 crc kubenswrapper[4726]: E1004 04:41:19.503340 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:41:20 crc kubenswrapper[4726]: I1004 04:41:20.516599 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b492feec-89af-4e54-b0e6-012436301beb" path="/var/lib/kubelet/pods/b492feec-89af-4e54-b0e6-012436301beb/volumes" Oct 04 04:41:30 crc kubenswrapper[4726]: I1004 04:41:30.503294 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:41:30 crc kubenswrapper[4726]: E1004 04:41:30.504137 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:41:43 crc kubenswrapper[4726]: I1004 04:41:43.502234 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:41:43 crc kubenswrapper[4726]: E1004 04:41:43.503007 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:41:56 crc kubenswrapper[4726]: I1004 04:41:56.508334 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:41:56 crc kubenswrapper[4726]: E1004 04:41:56.509538 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:42:09 crc kubenswrapper[4726]: I1004 04:42:09.503124 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:42:09 crc kubenswrapper[4726]: E1004 04:42:09.503936 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:42:20 crc kubenswrapper[4726]: I1004 04:42:20.502719 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:42:20 crc kubenswrapper[4726]: E1004 04:42:20.503510 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:42:34 crc kubenswrapper[4726]: I1004 04:42:34.503190 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:42:34 crc kubenswrapper[4726]: E1004 04:42:34.503845 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:42:48 crc kubenswrapper[4726]: I1004 04:42:48.502821 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:42:48 crc kubenswrapper[4726]: E1004 04:42:48.503541 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:42:53 crc kubenswrapper[4726]: I1004 04:42:53.385283 4726 scope.go:117] "RemoveContainer" containerID="33eb50e92869fe10a60b422e60f3ea9777c2e89394fcb239b0cdaa747aabc805" Oct 04 04:42:53 crc kubenswrapper[4726]: I1004 04:42:53.407698 4726 scope.go:117] "RemoveContainer" containerID="661bb426475da1a1fb0f6d7468c9a18dc25c301565609099cf65af8dde98820a" Oct 04 04:42:53 crc kubenswrapper[4726]: I1004 04:42:53.453091 4726 scope.go:117] "RemoveContainer" containerID="0e814e7a35be3d228208aa9cd093798c5cdfbdf5fc355ff0a7b09470293310d0" Oct 04 04:43:03 crc kubenswrapper[4726]: I1004 04:43:03.502527 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:43:03 crc kubenswrapper[4726]: E1004 04:43:03.503487 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:43:14 crc kubenswrapper[4726]: I1004 04:43:14.502395 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:43:14 crc kubenswrapper[4726]: E1004 04:43:14.503301 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:43:27 crc kubenswrapper[4726]: I1004 04:43:27.502993 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:43:27 crc kubenswrapper[4726]: E1004 04:43:27.503988 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:43:39 crc kubenswrapper[4726]: I1004 04:43:39.502639 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:43:39 crc kubenswrapper[4726]: E1004 04:43:39.503744 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:43:53 crc kubenswrapper[4726]: I1004 04:43:53.502388 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:43:53 crc kubenswrapper[4726]: E1004 04:43:53.503274 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:44:05 crc kubenswrapper[4726]: I1004 04:44:05.502896 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:44:05 crc kubenswrapper[4726]: E1004 04:44:05.503855 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:44:18 crc kubenswrapper[4726]: I1004 04:44:18.502989 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:44:18 crc kubenswrapper[4726]: E1004 04:44:18.503836 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:44:33 crc kubenswrapper[4726]: I1004 04:44:33.502814 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:44:33 crc kubenswrapper[4726]: E1004 04:44:33.503654 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:44:47 crc kubenswrapper[4726]: I1004 04:44:47.502650 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:44:47 crc kubenswrapper[4726]: E1004 04:44:47.503595 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:44:59 crc kubenswrapper[4726]: I1004 04:44:59.502161 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:44:59 crc kubenswrapper[4726]: E1004 04:44:59.502867 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142387 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf"] Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142775 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="extract-utilities" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142788 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="extract-utilities" Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142804 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="extract-content" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142810 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="extract-content" Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142827 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142833 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142846 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142852 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142882 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="extract-utilities" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142890 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="extract-utilities" Oct 04 04:45:00 crc kubenswrapper[4726]: E1004 04:45:00.142903 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="extract-content" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.142909 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="extract-content" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.143144 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="b492feec-89af-4e54-b0e6-012436301beb" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.143176 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="840b686c-580f-44aa-9d58-aff48accfbf1" containerName="registry-server" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.143825 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.146227 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.146602 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.158588 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf"] Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.244282 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9phxp\" (UniqueName: \"kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.244394 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.244600 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.346610 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9phxp\" (UniqueName: \"kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.346661 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.346703 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.347582 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.354268 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.364737 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9phxp\" (UniqueName: \"kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp\") pod \"collect-profiles-29325885-nrklf\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.513260 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:00 crc kubenswrapper[4726]: I1004 04:45:00.978819 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf"] Oct 04 04:45:01 crc kubenswrapper[4726]: I1004 04:45:01.455954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" event={"ID":"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc","Type":"ContainerStarted","Data":"33a7f4f1302b371d392e4ed6c4e1ad2812482f28b550a7ef2dc8798155d139e0"} Oct 04 04:45:01 crc kubenswrapper[4726]: I1004 04:45:01.456563 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" event={"ID":"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc","Type":"ContainerStarted","Data":"7c5f47d4fb6da4feb0598260e3ea669b8ee404a10f9027cd21b4af589fc894f0"} Oct 04 04:45:01 crc kubenswrapper[4726]: I1004 04:45:01.488354 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" podStartSLOduration=1.488327981 podStartE2EDuration="1.488327981s" podCreationTimestamp="2025-10-04 04:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:45:01.479956397 +0000 UTC m=+3875.654579630" watchObservedRunningTime="2025-10-04 04:45:01.488327981 +0000 UTC m=+3875.662951204" Oct 04 04:45:02 crc kubenswrapper[4726]: I1004 04:45:02.465866 4726 generic.go:334] "Generic (PLEG): container finished" podID="e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" containerID="33a7f4f1302b371d392e4ed6c4e1ad2812482f28b550a7ef2dc8798155d139e0" exitCode=0 Oct 04 04:45:02 crc kubenswrapper[4726]: I1004 04:45:02.465941 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" event={"ID":"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc","Type":"ContainerDied","Data":"33a7f4f1302b371d392e4ed6c4e1ad2812482f28b550a7ef2dc8798155d139e0"} Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.864069 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.911008 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume\") pod \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.911224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9phxp\" (UniqueName: \"kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp\") pod \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.911284 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume\") pod \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\" (UID: \"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc\") " Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.911944 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" (UID: "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.921698 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" (UID: "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4726]: I1004 04:45:03.929349 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp" (OuterVolumeSpecName: "kube-api-access-9phxp") pod "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" (UID: "e88ede0e-44a3-49c9-91f2-8e871fa5f8cc"). InnerVolumeSpecName "kube-api-access-9phxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.013969 4726 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.014010 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9phxp\" (UniqueName: \"kubernetes.io/projected/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-kube-api-access-9phxp\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.014024 4726 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e88ede0e-44a3-49c9-91f2-8e871fa5f8cc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.483684 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" event={"ID":"e88ede0e-44a3-49c9-91f2-8e871fa5f8cc","Type":"ContainerDied","Data":"7c5f47d4fb6da4feb0598260e3ea669b8ee404a10f9027cd21b4af589fc894f0"} Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.483990 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c5f47d4fb6da4feb0598260e3ea669b8ee404a10f9027cd21b4af589fc894f0" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.483753 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-nrklf" Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.945464 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6"] Oct 04 04:45:04 crc kubenswrapper[4726]: I1004 04:45:04.953420 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325840-nc6h6"] Oct 04 04:45:06 crc kubenswrapper[4726]: I1004 04:45:06.521612 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6d2aabc-dbc5-4442-af4e-2515c2d7e19b" path="/var/lib/kubelet/pods/e6d2aabc-dbc5-4442-af4e-2515c2d7e19b/volumes" Oct 04 04:45:11 crc kubenswrapper[4726]: I1004 04:45:11.503250 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:45:11 crc kubenswrapper[4726]: E1004 04:45:11.504288 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:45:13 crc kubenswrapper[4726]: I1004 04:45:13.567405 4726 generic.go:334] "Generic (PLEG): container finished" podID="de0aaf35-392c-4705-b554-1d4768c5cff1" containerID="16770c96268e50fcfc974cbbf6862eee0beccb42481a2fee2f994fa7cadd29f0" exitCode=0 Oct 04 04:45:13 crc kubenswrapper[4726]: I1004 04:45:13.567523 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de0aaf35-392c-4705-b554-1d4768c5cff1","Type":"ContainerDied","Data":"16770c96268e50fcfc974cbbf6862eee0beccb42481a2fee2f994fa7cadd29f0"} Oct 04 04:45:14 crc kubenswrapper[4726]: I1004 04:45:14.914679 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.030768 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.030878 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.030917 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.030952 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.031091 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.031197 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.031231 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.031267 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2jhs\" (UniqueName: \"kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.031322 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data\") pod \"de0aaf35-392c-4705-b554-1d4768c5cff1\" (UID: \"de0aaf35-392c-4705-b554-1d4768c5cff1\") " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.032544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.032685 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data" (OuterVolumeSpecName: "config-data") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.040284 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.053544 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "test-operator-logs") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.053578 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs" (OuterVolumeSpecName: "kube-api-access-n2jhs") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "kube-api-access-n2jhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.060820 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.060926 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.071871 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.091327 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "de0aaf35-392c-4705-b554-1d4768c5cff1" (UID: "de0aaf35-392c-4705-b554-1d4768c5cff1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133470 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133507 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133518 4726 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/de0aaf35-392c-4705-b554-1d4768c5cff1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133529 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2jhs\" (UniqueName: \"kubernetes.io/projected/de0aaf35-392c-4705-b554-1d4768c5cff1-kube-api-access-n2jhs\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133539 4726 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de0aaf35-392c-4705-b554-1d4768c5cff1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133548 4726 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133556 4726 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133564 4726 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de0aaf35-392c-4705-b554-1d4768c5cff1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.133601 4726 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.155055 4726 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.235293 4726 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.589011 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"de0aaf35-392c-4705-b554-1d4768c5cff1","Type":"ContainerDied","Data":"e8b53caaec78e26a806c8f0e218f57da8a24e6feed7d6ba14458b6a89036cd07"} Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.589438 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8b53caaec78e26a806c8f0e218f57da8a24e6feed7d6ba14458b6a89036cd07" Oct 04 04:45:15 crc kubenswrapper[4726]: I1004 04:45:15.589089 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.179516 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 04:45:24 crc kubenswrapper[4726]: E1004 04:45:24.181077 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" containerName="collect-profiles" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.181142 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" containerName="collect-profiles" Oct 04 04:45:24 crc kubenswrapper[4726]: E1004 04:45:24.181210 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de0aaf35-392c-4705-b554-1d4768c5cff1" containerName="tempest-tests-tempest-tests-runner" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.181230 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="de0aaf35-392c-4705-b554-1d4768c5cff1" containerName="tempest-tests-tempest-tests-runner" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.181688 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="e88ede0e-44a3-49c9-91f2-8e871fa5f8cc" containerName="collect-profiles" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.181732 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="de0aaf35-392c-4705-b554-1d4768c5cff1" containerName="tempest-tests-tempest-tests-runner" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.182880 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.185734 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-nkvlq" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.193777 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.312326 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9sdn\" (UniqueName: \"kubernetes.io/projected/c20315a1-0428-4211-b771-49a5504712dd-kube-api-access-t9sdn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.312639 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.414999 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9sdn\" (UniqueName: \"kubernetes.io/projected/c20315a1-0428-4211-b771-49a5504712dd-kube-api-access-t9sdn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.415419 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.415847 4726 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.437310 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9sdn\" (UniqueName: \"kubernetes.io/projected/c20315a1-0428-4211-b771-49a5504712dd-kube-api-access-t9sdn\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.458251 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c20315a1-0428-4211-b771-49a5504712dd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.515948 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 04:45:24 crc kubenswrapper[4726]: I1004 04:45:24.961952 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 04:45:25 crc kubenswrapper[4726]: I1004 04:45:25.502361 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:45:25 crc kubenswrapper[4726]: E1004 04:45:25.503077 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:45:25 crc kubenswrapper[4726]: I1004 04:45:25.689868 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c20315a1-0428-4211-b771-49a5504712dd","Type":"ContainerStarted","Data":"083d9d72b33004fe4b4f3185ed097f217c771ce1d4e9bae60c27e83233e6f5eb"} Oct 04 04:45:26 crc kubenswrapper[4726]: I1004 04:45:26.701356 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c20315a1-0428-4211-b771-49a5504712dd","Type":"ContainerStarted","Data":"3854a79461c96966d3994fcf2ae5c75fe778d36dff6c478b08c1b3f111b9fd29"} Oct 04 04:45:26 crc kubenswrapper[4726]: I1004 04:45:26.728807 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9275260159999998 podStartE2EDuration="2.728778223s" podCreationTimestamp="2025-10-04 04:45:24 +0000 UTC" firstStartedPulling="2025-10-04 04:45:24.967101965 +0000 UTC m=+3899.141725178" lastFinishedPulling="2025-10-04 04:45:25.768354172 +0000 UTC m=+3899.942977385" observedRunningTime="2025-10-04 04:45:26.715205886 +0000 UTC m=+3900.889829109" watchObservedRunningTime="2025-10-04 04:45:26.728778223 +0000 UTC m=+3900.903401436" Oct 04 04:45:39 crc kubenswrapper[4726]: I1004 04:45:39.502496 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:45:39 crc kubenswrapper[4726]: E1004 04:45:39.503345 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.930693 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s7tm6/must-gather-7fvw9"] Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.932913 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.935078 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-s7tm6"/"default-dockercfg-sttml" Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.935230 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s7tm6"/"openshift-service-ca.crt" Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.935461 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-s7tm6"/"kube-root-ca.crt" Oct 04 04:45:43 crc kubenswrapper[4726]: I1004 04:45:43.941188 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s7tm6/must-gather-7fvw9"] Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.117090 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cw4d\" (UniqueName: \"kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.117794 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.219435 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.219493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cw4d\" (UniqueName: \"kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.220009 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.239445 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cw4d\" (UniqueName: \"kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d\") pod \"must-gather-7fvw9\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.262706 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.726959 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-s7tm6/must-gather-7fvw9"] Oct 04 04:45:44 crc kubenswrapper[4726]: I1004 04:45:44.862903 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" event={"ID":"566782cd-404b-4bb9-9b0f-7d3892e15c0e","Type":"ContainerStarted","Data":"0fec6894eeaa6d87122713eba8a9e0b67461d1e2d455fcb2c7dbe8906f164d72"} Oct 04 04:45:51 crc kubenswrapper[4726]: I1004 04:45:51.936093 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" event={"ID":"566782cd-404b-4bb9-9b0f-7d3892e15c0e","Type":"ContainerStarted","Data":"8558bae34980123ba2f6c197b667a27b5a1ec7e1110ef3dd132f9e9cc1ff2408"} Oct 04 04:45:51 crc kubenswrapper[4726]: I1004 04:45:51.936662 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" event={"ID":"566782cd-404b-4bb9-9b0f-7d3892e15c0e","Type":"ContainerStarted","Data":"9075792a1a4d2f84a04b7de3665ec39cd74712ee6d7fbe5b0ff6c1f09659fe8b"} Oct 04 04:45:51 crc kubenswrapper[4726]: I1004 04:45:51.952660 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" podStartSLOduration=2.91892669 podStartE2EDuration="8.952638602s" podCreationTimestamp="2025-10-04 04:45:43 +0000 UTC" firstStartedPulling="2025-10-04 04:45:44.731837621 +0000 UTC m=+3918.906460834" lastFinishedPulling="2025-10-04 04:45:50.765549533 +0000 UTC m=+3924.940172746" observedRunningTime="2025-10-04 04:45:51.951651627 +0000 UTC m=+3926.126274830" watchObservedRunningTime="2025-10-04 04:45:51.952638602 +0000 UTC m=+3926.127261815" Oct 04 04:45:52 crc kubenswrapper[4726]: I1004 04:45:52.501873 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:45:52 crc kubenswrapper[4726]: E1004 04:45:52.502264 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:45:53 crc kubenswrapper[4726]: I1004 04:45:53.561140 4726 scope.go:117] "RemoveContainer" containerID="b0251c200d8fb3d2d318b3cf1fdbe0df206b5904426ba80a9dad30b373391f77" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.480931 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-6mp2l"] Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.482792 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.571783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.571873 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw5gp\" (UniqueName: \"kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.673544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.673633 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw5gp\" (UniqueName: \"kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.673866 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.690875 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw5gp\" (UniqueName: \"kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp\") pod \"crc-debug-6mp2l\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.800714 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:45:56 crc kubenswrapper[4726]: W1004 04:45:56.842894 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedfe23c5_b465_4805_9a29_1db18915d796.slice/crio-d35088865aad32d232cc648448b197893fb09687dcf8d7e9710b1bcba87dce3c WatchSource:0}: Error finding container d35088865aad32d232cc648448b197893fb09687dcf8d7e9710b1bcba87dce3c: Status 404 returned error can't find the container with id d35088865aad32d232cc648448b197893fb09687dcf8d7e9710b1bcba87dce3c Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.845478 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:45:56 crc kubenswrapper[4726]: I1004 04:45:56.983332 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" event={"ID":"edfe23c5-b465-4805-9a29-1db18915d796","Type":"ContainerStarted","Data":"d35088865aad32d232cc648448b197893fb09687dcf8d7e9710b1bcba87dce3c"} Oct 04 04:46:03 crc kubenswrapper[4726]: I1004 04:46:03.502801 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:46:03 crc kubenswrapper[4726]: E1004 04:46:03.503855 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:46:09 crc kubenswrapper[4726]: I1004 04:46:09.129674 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" event={"ID":"edfe23c5-b465-4805-9a29-1db18915d796","Type":"ContainerStarted","Data":"54d4ab076a0f3510889bcc7e6485e84a2f4f77b6bef96b4ecf71fab7d00aecce"} Oct 04 04:46:09 crc kubenswrapper[4726]: I1004 04:46:09.155474 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" podStartSLOduration=1.912056109 podStartE2EDuration="13.155447756s" podCreationTimestamp="2025-10-04 04:45:56 +0000 UTC" firstStartedPulling="2025-10-04 04:45:56.845073977 +0000 UTC m=+3931.019697200" lastFinishedPulling="2025-10-04 04:46:08.088465634 +0000 UTC m=+3942.263088847" observedRunningTime="2025-10-04 04:46:09.147549459 +0000 UTC m=+3943.322172672" watchObservedRunningTime="2025-10-04 04:46:09.155447756 +0000 UTC m=+3943.330070989" Oct 04 04:46:18 crc kubenswrapper[4726]: I1004 04:46:18.502825 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:46:19 crc kubenswrapper[4726]: I1004 04:46:19.221850 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882"} Oct 04 04:46:54 crc kubenswrapper[4726]: I1004 04:46:54.992701 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:46:54 crc kubenswrapper[4726]: I1004 04:46:54.995136 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.011697 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.064702 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq59n\" (UniqueName: \"kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.064873 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.064961 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.167213 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.167597 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.167664 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq59n\" (UniqueName: \"kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.168165 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.168325 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.189034 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq59n\" (UniqueName: \"kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n\") pod \"community-operators-sd5rh\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:55 crc kubenswrapper[4726]: I1004 04:46:55.313526 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:46:56 crc kubenswrapper[4726]: I1004 04:46:56.760446 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.383715 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.385948 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.401817 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.420538 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.420618 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2297c\" (UniqueName: \"kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.420697 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.522844 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2297c\" (UniqueName: \"kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.523325 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.523544 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.523936 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.524089 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.543068 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2297c\" (UniqueName: \"kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c\") pod \"redhat-operators-h2rxr\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.651951 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerID="25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace" exitCode=0 Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.652001 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerDied","Data":"25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace"} Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.652033 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerStarted","Data":"18106f65d6823b8c1518fb5b6ff2c2cd9d1cb54675d7b045b673fff0c3bed8a7"} Oct 04 04:46:57 crc kubenswrapper[4726]: I1004 04:46:57.721669 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:46:58 crc kubenswrapper[4726]: I1004 04:46:58.274303 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:46:58 crc kubenswrapper[4726]: W1004 04:46:58.520504 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6464d1c7_5a26_4c4e_beed_e9037e368648.slice/crio-ff0d7c311a175cc49e9d94b31cdfb2f24fe77dcaeda28d22dc06417c4726c286 WatchSource:0}: Error finding container ff0d7c311a175cc49e9d94b31cdfb2f24fe77dcaeda28d22dc06417c4726c286: Status 404 returned error can't find the container with id ff0d7c311a175cc49e9d94b31cdfb2f24fe77dcaeda28d22dc06417c4726c286 Oct 04 04:46:58 crc kubenswrapper[4726]: I1004 04:46:58.664956 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerStarted","Data":"ff0d7c311a175cc49e9d94b31cdfb2f24fe77dcaeda28d22dc06417c4726c286"} Oct 04 04:46:59 crc kubenswrapper[4726]: I1004 04:46:59.681493 4726 generic.go:334] "Generic (PLEG): container finished" podID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerID="2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a" exitCode=0 Oct 04 04:46:59 crc kubenswrapper[4726]: I1004 04:46:59.681601 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerDied","Data":"2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a"} Oct 04 04:46:59 crc kubenswrapper[4726]: I1004 04:46:59.683926 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerID="3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29" exitCode=0 Oct 04 04:46:59 crc kubenswrapper[4726]: I1004 04:46:59.683954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerDied","Data":"3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29"} Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.036341 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56bb546-6dhls_d74c5864-48ad-4376-b318-071c1925ad4c/barbican-api/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.130152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56bb546-6dhls_d74c5864-48ad-4376-b318-071c1925ad4c/barbican-api-log/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.208406 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fccfff78b-bs7hd_0f684dad-f165-4648-a5b0-e1e94b2cdc44/barbican-keystone-listener/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.337162 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fccfff78b-bs7hd_0f684dad-f165-4648-a5b0-e1e94b2cdc44/barbican-keystone-listener-log/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.397379 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78dd78b9bc-pwdzf_be9d885c-08ac-43eb-a1a7-2829719d4187/barbican-worker/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.550223 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78dd78b9bc-pwdzf_be9d885c-08ac-43eb-a1a7-2829719d4187/barbican-worker-log/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.649733 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg_aa6f54c4-9900-41b2-a047-6dda8304ff99/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.698062 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerStarted","Data":"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6"} Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.717299 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sd5rh" podStartSLOduration=3.929504708 podStartE2EDuration="6.717280659s" podCreationTimestamp="2025-10-04 04:46:54 +0000 UTC" firstStartedPulling="2025-10-04 04:46:57.654496744 +0000 UTC m=+3991.829119957" lastFinishedPulling="2025-10-04 04:47:00.442272695 +0000 UTC m=+3994.616895908" observedRunningTime="2025-10-04 04:47:00.714989709 +0000 UTC m=+3994.889612922" watchObservedRunningTime="2025-10-04 04:47:00.717280659 +0000 UTC m=+3994.891903872" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.878052 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/ceilometer-notification-agent/0.log" Oct 04 04:47:00 crc kubenswrapper[4726]: I1004 04:47:00.904551 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/ceilometer-central-agent/0.log" Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.151489 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/proxy-httpd/0.log" Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.228752 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/sg-core/0.log" Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.446454 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cc91ace6-aab5-4a5c-9549-131d3303f5d5/cinder-api/0.log" Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.527133 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cc91ace6-aab5-4a5c-9549-131d3303f5d5/cinder-api-log/0.log" Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.707657 4726 generic.go:334] "Generic (PLEG): container finished" podID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerID="1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7" exitCode=0 Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.707968 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerDied","Data":"1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7"} Oct 04 04:47:01 crc kubenswrapper[4726]: I1004 04:47:01.820390 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2105d235-5b6d-40ea-80d4-0f2033e0bf5d/cinder-scheduler/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.015940 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2105d235-5b6d-40ea-80d4-0f2033e0bf5d/probe/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.245435 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7_ebf88812-a311-4c5a-8698-d50f551446f2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.302860 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q_13859a35-3c01-4314-a416-0d3d04828cdc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.540642 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5_dec26676-436d-4d9a-8337-0c9ac86c6168/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.697732 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/init/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.939251 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/init/0.log" Oct 04 04:47:02 crc kubenswrapper[4726]: I1004 04:47:02.984568 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/dnsmasq-dns/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.023174 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5_4a4f5e93-bd73-42f9-8282-39e559ad08a7/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.164393 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bae4a133-24a9-4ba4-8050-5e1a15bc8ed3/glance-httpd/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.192121 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bae4a133-24a9-4ba4-8050-5e1a15bc8ed3/glance-log/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.375718 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_636ad703-b66b-42f6-aca7-00900d0fad9b/glance-log/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.454129 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_636ad703-b66b-42f6-aca7-00900d0fad9b/glance-httpd/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.575234 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7786888bd-6khpf_888736e3-9e3d-402b-9ea3-79a0ba740fe7/horizon/0.log" Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.727270 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerStarted","Data":"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467"} Oct 04 04:47:03 crc kubenswrapper[4726]: I1004 04:47:03.772188 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nzrft_70435c0c-e859-49bd-9e2e-624f66c3be54/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.012898 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7786888bd-6khpf_888736e3-9e3d-402b-9ea3-79a0ba740fe7/horizon-log/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.024896 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-v85jc_1faabfae-0c50-425d-af5b-529a69ee6791/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.217525 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-79b54b49b-mx2kb_cb8197e3-5f74-443a-bf22-78a28365cfaf/keystone-api/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.256217 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325841-fnfsx_db67cb41-86ca-41f3-b012-1bfc33e84a37/keystone-cron/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.385822 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c4db3f45-a9c5-4092-a236-f3c345e2303d/kube-state-metrics/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.554907 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d_836b7c28-6c4b-44c1-9f6f-ff52d2d7b107/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.758759 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h2rxr" podStartSLOduration=5.050915596 podStartE2EDuration="7.758744881s" podCreationTimestamp="2025-10-04 04:46:57 +0000 UTC" firstStartedPulling="2025-10-04 04:46:59.722307784 +0000 UTC m=+3993.896930997" lastFinishedPulling="2025-10-04 04:47:02.430137069 +0000 UTC m=+3996.604760282" observedRunningTime="2025-10-04 04:47:04.757956401 +0000 UTC m=+3998.932579614" watchObservedRunningTime="2025-10-04 04:47:04.758744881 +0000 UTC m=+3998.933368094" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.951237 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5bd64b8469-btvtm_d97c6241-e072-4012-a61e-c5e855407e99/neutron-httpd/0.log" Oct 04 04:47:04 crc kubenswrapper[4726]: I1004 04:47:04.976484 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5bd64b8469-btvtm_d97c6241-e072-4012-a61e-c5e855407e99/neutron-api/0.log" Oct 04 04:47:05 crc kubenswrapper[4726]: I1004 04:47:05.314194 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:05 crc kubenswrapper[4726]: I1004 04:47:05.314320 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:05 crc kubenswrapper[4726]: I1004 04:47:05.541678 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz_cbf08126-f813-40fe-b451-71cab2c2fd8a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:06 crc kubenswrapper[4726]: I1004 04:47:06.204122 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5c95d78d-1f21-48f0-903e-713cde259066/nova-api-log/0.log" Oct 04 04:47:06 crc kubenswrapper[4726]: I1004 04:47:06.347455 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_76642298-36b4-4bde-ba49-3b0a2e56494d/nova-cell0-conductor-conductor/0.log" Oct 04 04:47:06 crc kubenswrapper[4726]: I1004 04:47:06.400219 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-sd5rh" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="registry-server" probeResult="failure" output=< Oct 04 04:47:06 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:47:06 crc kubenswrapper[4726]: > Oct 04 04:47:06 crc kubenswrapper[4726]: I1004 04:47:06.460609 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5c95d78d-1f21-48f0-903e-713cde259066/nova-api-api/0.log" Oct 04 04:47:06 crc kubenswrapper[4726]: I1004 04:47:06.733087 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_11f9ad5d-1851-40d1-b8f9-cfd476b498ce/nova-cell1-conductor-conductor/0.log" Oct 04 04:47:07 crc kubenswrapper[4726]: I1004 04:47:07.232273 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_df4fee1d-47f1-48e0-98bd-f4db94f5a543/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 04:47:07 crc kubenswrapper[4726]: I1004 04:47:07.427747 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-j6dh8_cb83849c-9bbd-479b-9a49-e99ea5247ab1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:07 crc kubenswrapper[4726]: I1004 04:47:07.603654 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f13faed-0be2-4916-80ab-f9d7c7d196a3/nova-metadata-log/0.log" Oct 04 04:47:07 crc kubenswrapper[4726]: I1004 04:47:07.722380 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:07 crc kubenswrapper[4726]: I1004 04:47:07.723869 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.129091 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2a84708a-3dbc-477f-b151-d68b83b4617b/nova-scheduler-scheduler/0.log" Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.357614 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/mysql-bootstrap/0.log" Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.692462 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/mysql-bootstrap/0.log" Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.707453 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/galera/0.log" Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.773411 4726 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h2rxr" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="registry-server" probeResult="failure" output=< Oct 04 04:47:08 crc kubenswrapper[4726]: timeout: failed to connect service ":50051" within 1s Oct 04 04:47:08 crc kubenswrapper[4726]: > Oct 04 04:47:08 crc kubenswrapper[4726]: I1004 04:47:08.985420 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/mysql-bootstrap/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.166821 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f13faed-0be2-4916-80ab-f9d7c7d196a3/nova-metadata-metadata/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.180992 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/galera/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.191565 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/mysql-bootstrap/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.416794 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f592dceb-e227-48ea-a6d6-0b9513ec5b46/openstackclient/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.653158 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5kwdw_8675d591-017c-4479-bcb2-dbb70c2f9f93/ovn-controller/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.674122 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r526s_507a6d8a-447d-436c-918d-7e2f447705bf/openstack-network-exporter/0.log" Oct 04 04:47:09 crc kubenswrapper[4726]: I1004 04:47:09.933233 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server-init/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.105632 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovs-vswitchd/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.118808 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.139064 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server-init/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.316165 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jfxbp_e1087fbc-edc3-47d2-8229-fb1654ce31eb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.514334 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f3ae4630-e488-43db-a57a-762c75de75aa/openstack-network-exporter/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.593368 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f3ae4630-e488-43db-a57a-762c75de75aa/ovn-northd/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.750099 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_164dc913-a191-4c88-ad58-e2a08341d0a3/openstack-network-exporter/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.829882 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_164dc913-a191-4c88-ad58-e2a08341d0a3/ovsdbserver-nb/0.log" Oct 04 04:47:10 crc kubenswrapper[4726]: I1004 04:47:10.987468 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d298c1d3-002d-4847-bddc-63df803a9029/openstack-network-exporter/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.054340 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d298c1d3-002d-4847-bddc-63df803a9029/ovsdbserver-sb/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.316905 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f697b85fb-mcpgk_c10188ee-0fc9-4985-a21c-88fbc85e1e3c/placement-api/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.387360 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f697b85fb-mcpgk_c10188ee-0fc9-4985-a21c-88fbc85e1e3c/placement-log/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.511183 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/setup-container/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.707781 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/setup-container/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.750720 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/rabbitmq/0.log" Oct 04 04:47:11 crc kubenswrapper[4726]: I1004 04:47:11.930428 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/setup-container/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.114990 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/setup-container/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.138199 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/rabbitmq/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.340605 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-w8969_0c4effe5-c8f2-4f82-b5b0-3052858642f1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.440755 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6mqds_2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.648884 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2_db0467b4-3329-4278-b094-02b70e240f01/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:12 crc kubenswrapper[4726]: I1004 04:47:12.850547 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-786b8_9465de2d-b388-410c-90da-31beba8b5dbe/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.085934 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mjmvr_e0ad6b26-b887-46ea-9402-1e9bf3a77de9/ssh-known-hosts-edpm-deployment/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.265329 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bdb856b6c-f5xfb_503e5303-a481-4184-9bb0-2369123d5267/proxy-server/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.329152 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bdb856b6c-f5xfb_503e5303-a481-4184-9bb0-2369123d5267/proxy-httpd/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.437157 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qxprg_63b2f4dc-088d-47ba-aad0-e0de06d93828/swift-ring-rebalance/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.584584 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-auditor/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.628848 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-reaper/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.784766 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-server/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.804901 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-replicator/0.log" Oct 04 04:47:13 crc kubenswrapper[4726]: I1004 04:47:13.843998 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-auditor/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.024019 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-replicator/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.041620 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-server/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.106950 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-updater/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.803238 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-auditor/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.818233 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-expirer/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.830080 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-replicator/0.log" Oct 04 04:47:14 crc kubenswrapper[4726]: I1004 04:47:14.999146 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-updater/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.022124 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-server/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.154632 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/rsync/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.205931 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/swift-recon-cron/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.365485 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.419923 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wclx9_e9a93781-fcc5-4323-bb02-934800e8de8f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.420174 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.561630 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_de0aaf35-392c-4705-b554-1d4768c5cff1/tempest-tests-tempest-tests-runner/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.599278 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.711902 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c20315a1-0428-4211-b771-49a5504712dd/test-operator-logs-container/0.log" Oct 04 04:47:15 crc kubenswrapper[4726]: I1004 04:47:15.869998 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w_affede2f-dbd9-44c4-ae15-af1fb96ce9f6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:47:16 crc kubenswrapper[4726]: I1004 04:47:16.862331 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sd5rh" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="registry-server" containerID="cri-o://38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6" gracePeriod=2 Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.397885 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.534540 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content\") pod \"2dfac1ba-7047-4655-80e8-51d8d9b66067\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.534601 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities\") pod \"2dfac1ba-7047-4655-80e8-51d8d9b66067\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.534744 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq59n\" (UniqueName: \"kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n\") pod \"2dfac1ba-7047-4655-80e8-51d8d9b66067\" (UID: \"2dfac1ba-7047-4655-80e8-51d8d9b66067\") " Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.537295 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities" (OuterVolumeSpecName: "utilities") pod "2dfac1ba-7047-4655-80e8-51d8d9b66067" (UID: "2dfac1ba-7047-4655-80e8-51d8d9b66067"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.557863 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n" (OuterVolumeSpecName: "kube-api-access-dq59n") pod "2dfac1ba-7047-4655-80e8-51d8d9b66067" (UID: "2dfac1ba-7047-4655-80e8-51d8d9b66067"). InnerVolumeSpecName "kube-api-access-dq59n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.585702 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dfac1ba-7047-4655-80e8-51d8d9b66067" (UID: "2dfac1ba-7047-4655-80e8-51d8d9b66067"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.637060 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.637091 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq59n\" (UniqueName: \"kubernetes.io/projected/2dfac1ba-7047-4655-80e8-51d8d9b66067-kube-api-access-dq59n\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.637114 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dfac1ba-7047-4655-80e8-51d8d9b66067-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.776403 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.837094 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.878599 4726 generic.go:334] "Generic (PLEG): container finished" podID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerID="38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6" exitCode=0 Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.878683 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sd5rh" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.878704 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerDied","Data":"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6"} Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.878763 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sd5rh" event={"ID":"2dfac1ba-7047-4655-80e8-51d8d9b66067","Type":"ContainerDied","Data":"18106f65d6823b8c1518fb5b6ff2c2cd9d1cb54675d7b045b673fff0c3bed8a7"} Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.878789 4726 scope.go:117] "RemoveContainer" containerID="38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.908763 4726 scope.go:117] "RemoveContainer" containerID="3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.918425 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.931230 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sd5rh"] Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.960652 4726 scope.go:117] "RemoveContainer" containerID="25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.991461 4726 scope.go:117] "RemoveContainer" containerID="38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6" Oct 04 04:47:17 crc kubenswrapper[4726]: E1004 04:47:17.992880 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6\": container with ID starting with 38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6 not found: ID does not exist" containerID="38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.993369 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6"} err="failed to get container status \"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6\": rpc error: code = NotFound desc = could not find container \"38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6\": container with ID starting with 38d344ff67054cbb10aeb83f4957325e5398feb56cd87d103b8d9ebec01a17e6 not found: ID does not exist" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.994238 4726 scope.go:117] "RemoveContainer" containerID="3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29" Oct 04 04:47:17 crc kubenswrapper[4726]: E1004 04:47:17.997277 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29\": container with ID starting with 3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29 not found: ID does not exist" containerID="3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.997337 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29"} err="failed to get container status \"3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29\": rpc error: code = NotFound desc = could not find container \"3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29\": container with ID starting with 3bdd39b5e827c7af82d3499aa647900d2a223b7c9a38f78688c615187ef9ec29 not found: ID does not exist" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.997372 4726 scope.go:117] "RemoveContainer" containerID="25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace" Oct 04 04:47:17 crc kubenswrapper[4726]: E1004 04:47:17.997983 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace\": container with ID starting with 25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace not found: ID does not exist" containerID="25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace" Oct 04 04:47:17 crc kubenswrapper[4726]: I1004 04:47:17.998006 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace"} err="failed to get container status \"25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace\": rpc error: code = NotFound desc = could not find container \"25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace\": container with ID starting with 25e4226f250b710fcbc82306f080fdbd9d69391aec42a25a6d91890589a4cace not found: ID does not exist" Oct 04 04:47:18 crc kubenswrapper[4726]: I1004 04:47:18.514519 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" path="/var/lib/kubelet/pods/2dfac1ba-7047-4655-80e8-51d8d9b66067/volumes" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.001837 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.002682 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h2rxr" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="registry-server" containerID="cri-o://16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467" gracePeriod=2 Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.513421 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.597777 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities\") pod \"6464d1c7-5a26-4c4e-beed-e9037e368648\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.598030 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2297c\" (UniqueName: \"kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c\") pod \"6464d1c7-5a26-4c4e-beed-e9037e368648\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.598094 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content\") pod \"6464d1c7-5a26-4c4e-beed-e9037e368648\" (UID: \"6464d1c7-5a26-4c4e-beed-e9037e368648\") " Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.600497 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities" (OuterVolumeSpecName: "utilities") pod "6464d1c7-5a26-4c4e-beed-e9037e368648" (UID: "6464d1c7-5a26-4c4e-beed-e9037e368648"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.609282 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c" (OuterVolumeSpecName: "kube-api-access-2297c") pod "6464d1c7-5a26-4c4e-beed-e9037e368648" (UID: "6464d1c7-5a26-4c4e-beed-e9037e368648"). InnerVolumeSpecName "kube-api-access-2297c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.701118 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2297c\" (UniqueName: \"kubernetes.io/projected/6464d1c7-5a26-4c4e-beed-e9037e368648-kube-api-access-2297c\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.701163 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.711612 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6464d1c7-5a26-4c4e-beed-e9037e368648" (UID: "6464d1c7-5a26-4c4e-beed-e9037e368648"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.802634 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6464d1c7-5a26-4c4e-beed-e9037e368648-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.920537 4726 generic.go:334] "Generic (PLEG): container finished" podID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerID="16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467" exitCode=0 Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.920586 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerDied","Data":"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467"} Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.920623 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h2rxr" event={"ID":"6464d1c7-5a26-4c4e-beed-e9037e368648","Type":"ContainerDied","Data":"ff0d7c311a175cc49e9d94b31cdfb2f24fe77dcaeda28d22dc06417c4726c286"} Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.920643 4726 scope.go:117] "RemoveContainer" containerID="16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.920921 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h2rxr" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.943733 4726 scope.go:117] "RemoveContainer" containerID="1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7" Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.961316 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.973063 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h2rxr"] Oct 04 04:47:20 crc kubenswrapper[4726]: I1004 04:47:20.980031 4726 scope.go:117] "RemoveContainer" containerID="2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.030867 4726 scope.go:117] "RemoveContainer" containerID="16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467" Oct 04 04:47:21 crc kubenswrapper[4726]: E1004 04:47:21.031529 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467\": container with ID starting with 16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467 not found: ID does not exist" containerID="16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.031561 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467"} err="failed to get container status \"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467\": rpc error: code = NotFound desc = could not find container \"16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467\": container with ID starting with 16bb4afd98f39ccd942c5e2a87de3f4881f0a98fdbba313abc8bb006850c6467 not found: ID does not exist" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.031585 4726 scope.go:117] "RemoveContainer" containerID="1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7" Oct 04 04:47:21 crc kubenswrapper[4726]: E1004 04:47:21.031936 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7\": container with ID starting with 1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7 not found: ID does not exist" containerID="1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.031952 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7"} err="failed to get container status \"1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7\": rpc error: code = NotFound desc = could not find container \"1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7\": container with ID starting with 1a6027db74b83bb4e0e2139aae7ceb2f3d206d3319c3b4d0ccc12e9f178ac2c7 not found: ID does not exist" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.031965 4726 scope.go:117] "RemoveContainer" containerID="2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a" Oct 04 04:47:21 crc kubenswrapper[4726]: E1004 04:47:21.032247 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a\": container with ID starting with 2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a not found: ID does not exist" containerID="2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a" Oct 04 04:47:21 crc kubenswrapper[4726]: I1004 04:47:21.032274 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a"} err="failed to get container status \"2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a\": rpc error: code = NotFound desc = could not find container \"2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a\": container with ID starting with 2c7babd1859f3ada7fba1679d4e602bf0c37f2d3168dce01beea7765b362325a not found: ID does not exist" Oct 04 04:47:22 crc kubenswrapper[4726]: I1004 04:47:22.512895 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" path="/var/lib/kubelet/pods/6464d1c7-5a26-4c4e-beed-e9037e368648/volumes" Oct 04 04:47:23 crc kubenswrapper[4726]: I1004 04:47:23.612485 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5b1149f2-4d8f-4b1e-b5ec-b13f9833c706/memcached/0.log" Oct 04 04:48:09 crc kubenswrapper[4726]: I1004 04:48:09.436892 4726 generic.go:334] "Generic (PLEG): container finished" podID="edfe23c5-b465-4805-9a29-1db18915d796" containerID="54d4ab076a0f3510889bcc7e6485e84a2f4f77b6bef96b4ecf71fab7d00aecce" exitCode=0 Oct 04 04:48:09 crc kubenswrapper[4726]: I1004 04:48:09.437003 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" event={"ID":"edfe23c5-b465-4805-9a29-1db18915d796","Type":"ContainerDied","Data":"54d4ab076a0f3510889bcc7e6485e84a2f4f77b6bef96b4ecf71fab7d00aecce"} Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.568836 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.602184 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-6mp2l"] Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.609538 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-6mp2l"] Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.704782 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host\") pod \"edfe23c5-b465-4805-9a29-1db18915d796\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.704848 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw5gp\" (UniqueName: \"kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp\") pod \"edfe23c5-b465-4805-9a29-1db18915d796\" (UID: \"edfe23c5-b465-4805-9a29-1db18915d796\") " Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.704932 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host" (OuterVolumeSpecName: "host") pod "edfe23c5-b465-4805-9a29-1db18915d796" (UID: "edfe23c5-b465-4805-9a29-1db18915d796"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:48:10 crc kubenswrapper[4726]: I1004 04:48:10.705524 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/edfe23c5-b465-4805-9a29-1db18915d796-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:11 crc kubenswrapper[4726]: I1004 04:48:11.199351 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp" (OuterVolumeSpecName: "kube-api-access-tw5gp") pod "edfe23c5-b465-4805-9a29-1db18915d796" (UID: "edfe23c5-b465-4805-9a29-1db18915d796"). InnerVolumeSpecName "kube-api-access-tw5gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:11 crc kubenswrapper[4726]: I1004 04:48:11.216581 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw5gp\" (UniqueName: \"kubernetes.io/projected/edfe23c5-b465-4805-9a29-1db18915d796-kube-api-access-tw5gp\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:11 crc kubenswrapper[4726]: I1004 04:48:11.470136 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d35088865aad32d232cc648448b197893fb09687dcf8d7e9710b1bcba87dce3c" Oct 04 04:48:11 crc kubenswrapper[4726]: I1004 04:48:11.470207 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-6mp2l" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.477396 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-hkdmx"] Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478236 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="extract-content" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478255 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="extract-content" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478276 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="extract-utilities" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478283 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="extract-utilities" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478301 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478308 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478323 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="extract-utilities" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478331 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="extract-utilities" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478346 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edfe23c5-b465-4805-9a29-1db18915d796" containerName="container-00" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478355 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="edfe23c5-b465-4805-9a29-1db18915d796" containerName="container-00" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478385 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478392 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: E1004 04:48:12.478412 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="extract-content" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478419 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="extract-content" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478703 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="6464d1c7-5a26-4c4e-beed-e9037e368648" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478765 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dfac1ba-7047-4655-80e8-51d8d9b66067" containerName="registry-server" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.478777 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="edfe23c5-b465-4805-9a29-1db18915d796" containerName="container-00" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.479593 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.511903 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edfe23c5-b465-4805-9a29-1db18915d796" path="/var/lib/kubelet/pods/edfe23c5-b465-4805-9a29-1db18915d796/volumes" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.643777 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gmv4\" (UniqueName: \"kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.644202 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.746483 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gmv4\" (UniqueName: \"kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.746591 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:12 crc kubenswrapper[4726]: I1004 04:48:12.746790 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:13 crc kubenswrapper[4726]: I1004 04:48:13.098888 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gmv4\" (UniqueName: \"kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4\") pod \"crc-debug-hkdmx\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:13 crc kubenswrapper[4726]: I1004 04:48:13.101103 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:13 crc kubenswrapper[4726]: I1004 04:48:13.493221 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" event={"ID":"914f48f3-9e87-4134-818a-b52819ceb5fd","Type":"ContainerStarted","Data":"cc2f32e0b5fa919c33fa8ebf8fa4933b80e362193c6d2963f6201acbfa0bf60a"} Oct 04 04:48:14 crc kubenswrapper[4726]: I1004 04:48:14.507907 4726 generic.go:334] "Generic (PLEG): container finished" podID="914f48f3-9e87-4134-818a-b52819ceb5fd" containerID="b29db3b88ab435dd5b7cdc210cc024c5689fceca1b408eb03ce2c2e349959737" exitCode=0 Oct 04 04:48:14 crc kubenswrapper[4726]: I1004 04:48:14.521472 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" event={"ID":"914f48f3-9e87-4134-818a-b52819ceb5fd","Type":"ContainerDied","Data":"b29db3b88ab435dd5b7cdc210cc024c5689fceca1b408eb03ce2c2e349959737"} Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.628886 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.795006 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host\") pod \"914f48f3-9e87-4134-818a-b52819ceb5fd\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.795224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gmv4\" (UniqueName: \"kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4\") pod \"914f48f3-9e87-4134-818a-b52819ceb5fd\" (UID: \"914f48f3-9e87-4134-818a-b52819ceb5fd\") " Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.795222 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host" (OuterVolumeSpecName: "host") pod "914f48f3-9e87-4134-818a-b52819ceb5fd" (UID: "914f48f3-9e87-4134-818a-b52819ceb5fd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.795669 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/914f48f3-9e87-4134-818a-b52819ceb5fd-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.802454 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4" (OuterVolumeSpecName: "kube-api-access-8gmv4") pod "914f48f3-9e87-4134-818a-b52819ceb5fd" (UID: "914f48f3-9e87-4134-818a-b52819ceb5fd"). InnerVolumeSpecName "kube-api-access-8gmv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:15 crc kubenswrapper[4726]: I1004 04:48:15.899049 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gmv4\" (UniqueName: \"kubernetes.io/projected/914f48f3-9e87-4134-818a-b52819ceb5fd-kube-api-access-8gmv4\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:16 crc kubenswrapper[4726]: I1004 04:48:16.528137 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" Oct 04 04:48:16 crc kubenswrapper[4726]: I1004 04:48:16.528160 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-hkdmx" event={"ID":"914f48f3-9e87-4134-818a-b52819ceb5fd","Type":"ContainerDied","Data":"cc2f32e0b5fa919c33fa8ebf8fa4933b80e362193c6d2963f6201acbfa0bf60a"} Oct 04 04:48:16 crc kubenswrapper[4726]: I1004 04:48:16.528883 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc2f32e0b5fa919c33fa8ebf8fa4933b80e362193c6d2963f6201acbfa0bf60a" Oct 04 04:48:21 crc kubenswrapper[4726]: I1004 04:48:21.095740 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-hkdmx"] Oct 04 04:48:21 crc kubenswrapper[4726]: I1004 04:48:21.102907 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-hkdmx"] Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.256585 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-w4jgv"] Oct 04 04:48:22 crc kubenswrapper[4726]: E1004 04:48:22.257311 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="914f48f3-9e87-4134-818a-b52819ceb5fd" containerName="container-00" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.257327 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="914f48f3-9e87-4134-818a-b52819ceb5fd" containerName="container-00" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.257523 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="914f48f3-9e87-4134-818a-b52819ceb5fd" containerName="container-00" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.258125 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.402831 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.402888 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv2sw\" (UniqueName: \"kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.507730 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.507790 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv2sw\" (UniqueName: \"kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.508073 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.512813 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="914f48f3-9e87-4134-818a-b52819ceb5fd" path="/var/lib/kubelet/pods/914f48f3-9e87-4134-818a-b52819ceb5fd/volumes" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.528914 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv2sw\" (UniqueName: \"kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw\") pod \"crc-debug-w4jgv\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:22 crc kubenswrapper[4726]: I1004 04:48:22.577398 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:23 crc kubenswrapper[4726]: I1004 04:48:23.593945 4726 generic.go:334] "Generic (PLEG): container finished" podID="347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" containerID="e5d1a6e41e52f21b40ea1879b9a0af33ec958719f0f9f446a762e90be39da0b6" exitCode=0 Oct 04 04:48:23 crc kubenswrapper[4726]: I1004 04:48:23.594056 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" event={"ID":"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c","Type":"ContainerDied","Data":"e5d1a6e41e52f21b40ea1879b9a0af33ec958719f0f9f446a762e90be39da0b6"} Oct 04 04:48:23 crc kubenswrapper[4726]: I1004 04:48:23.594305 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" event={"ID":"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c","Type":"ContainerStarted","Data":"c59378b3a969cdfa3bbeadad49658907f0f8b13e00809800d249cd3a921ffdb8"} Oct 04 04:48:23 crc kubenswrapper[4726]: I1004 04:48:23.633205 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-w4jgv"] Oct 04 04:48:23 crc kubenswrapper[4726]: I1004 04:48:23.639705 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s7tm6/crc-debug-w4jgv"] Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.706639 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.852741 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host\") pod \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.852877 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host" (OuterVolumeSpecName: "host") pod "347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" (UID: "347b3f34-2be9-4bcf-803d-a67fc3cd7e3c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.852919 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv2sw\" (UniqueName: \"kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw\") pod \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\" (UID: \"347b3f34-2be9-4bcf-803d-a67fc3cd7e3c\") " Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.853611 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.858850 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw" (OuterVolumeSpecName: "kube-api-access-qv2sw") pod "347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" (UID: "347b3f34-2be9-4bcf-803d-a67fc3cd7e3c"). InnerVolumeSpecName "kube-api-access-qv2sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:24 crc kubenswrapper[4726]: I1004 04:48:24.955327 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv2sw\" (UniqueName: \"kubernetes.io/projected/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c-kube-api-access-qv2sw\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.618083 4726 scope.go:117] "RemoveContainer" containerID="e5d1a6e41e52f21b40ea1879b9a0af33ec958719f0f9f446a762e90be39da0b6" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.618098 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/crc-debug-w4jgv" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.647311 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.803446 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.830179 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.837725 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:48:25 crc kubenswrapper[4726]: I1004 04:48:25.995448 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.000344 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.016354 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/extract/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.187959 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-mfsl5_1e47df33-2621-4cfe-9c83-4d4bd1d616aa/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.253216 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-tzr27_f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.388532 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-mfsl5_1e47df33-2621-4cfe-9c83-4d4bd1d616aa/manager/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.394279 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-tzr27_f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3/manager/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.420681 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-gtshn_a520bf75-9223-4209-87d4-248ce3e0d190/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.523086 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" path="/var/lib/kubelet/pods/347b3f34-2be9-4bcf-803d-a67fc3cd7e3c/volumes" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.574839 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-gtshn_a520bf75-9223-4209-87d4-248ce3e0d190/manager/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.595160 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-cjx6d_79a719e7-5bce-424a-b837-ab0bca45936e/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.684539 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-cjx6d_79a719e7-5bce-424a-b837-ab0bca45936e/manager/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.788997 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-lqs6l_ac49881b-2d45-4ba0-8b60-8461a2d5f35b/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.791932 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-lqs6l_ac49881b-2d45-4ba0-8b60-8461a2d5f35b/manager/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.918718 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-fvw5p_5e5f2785-db94-4fdc-9cb2-10b1ce940694/kube-rbac-proxy/0.log" Oct 04 04:48:26 crc kubenswrapper[4726]: I1004 04:48:26.956913 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-fvw5p_5e5f2785-db94-4fdc-9cb2-10b1ce940694/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.060091 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7ddb876846-kgv8l_ea261e29-40c5-4a26-a02f-8a09e0c7320d/kube-rbac-proxy/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.173177 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-v72j6_9f398c24-0435-49f5-915c-a54efeadea0c/kube-rbac-proxy/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.289493 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-v72j6_9f398c24-0435-49f5-915c-a54efeadea0c/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.299032 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7ddb876846-kgv8l_ea261e29-40c5-4a26-a02f-8a09e0c7320d/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.397013 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-wwdj2_488fc85d-f45f-4d8d-93e2-b58ba9306a76/kube-rbac-proxy/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.536482 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-wwdj2_488fc85d-f45f-4d8d-93e2-b58ba9306a76/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.600376 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-6dpjp_6508ae63-4f75-496f-b7b6-5dcd8a0cb008/kube-rbac-proxy/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.623368 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-6dpjp_6508ae63-4f75-496f-b7b6-5dcd8a0cb008/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.776808 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-s67gj_d55e66e5-0c46-4df2-909c-fc739a26c5cc/kube-rbac-proxy/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.806549 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-s67gj_d55e66e5-0c46-4df2-909c-fc739a26c5cc/manager/0.log" Oct 04 04:48:27 crc kubenswrapper[4726]: I1004 04:48:27.946764 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lqsxr_d59f6f6b-419b-4ce3-adb6-11381aed6a5d/kube-rbac-proxy/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.030692 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lqsxr_d59f6f6b-419b-4ce3-adb6-11381aed6a5d/manager/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.064662 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mnndm_05b7da99-96e2-40be-8fc1-f847b8e7a643/kube-rbac-proxy/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.191760 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mnndm_05b7da99-96e2-40be-8fc1-f847b8e7a643/manager/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.246408 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-zkgst_cadfc4b1-9327-4a55-8544-44422572319c/kube-rbac-proxy/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.269730 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-zkgst_cadfc4b1-9327-4a55-8544-44422572319c/manager/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.387381 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt_6b245f13-4e4b-4a41-b148-482d581e2c5c/kube-rbac-proxy/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.423476 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt_6b245f13-4e4b-4a41-b148-482d581e2c5c/manager/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.584080 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-ff7c6f545-hgqmd_aaa6fd43-d6ed-4d86-9827-27d2529499e3/kube-rbac-proxy/0.log" Oct 04 04:48:28 crc kubenswrapper[4726]: I1004 04:48:28.731221 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5499555f6-blzgx_0bd87ab8-94f6-41a4-b76c-7774cac7b625/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.015228 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fzc75_40e39198-082e-4c71-b67c-6d845e45f782/registry-server/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.015239 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5499555f6-blzgx_0bd87ab8-94f6-41a4-b76c-7774cac7b625/operator/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.140910 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-8pnpc_e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.304575 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-lsxtw_efa34c5f-e531-4319-b5fb-7497c4d026f0/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.328630 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-8pnpc_e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828/manager/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.506339 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-lsxtw_efa34c5f-e531-4319-b5fb-7497c4d026f0/manager/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.536318 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7_962ba8ee-fbcf-4ef5-b740-bd4a782d0881/operator/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.667971 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-ff7c6f545-hgqmd_aaa6fd43-d6ed-4d86-9827-27d2529499e3/manager/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.672277 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hv9lz_5cb54ad2-85db-4291-87dc-4f0b9ea688c0/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.723598 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hv9lz_5cb54ad2-85db-4291-87dc-4f0b9ea688c0/manager/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.778365 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-2j9rb_42949778-256b-4809-b2ce-d633428e649c/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.900044 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4v7k8_392e7503-f32e-4e8f-ac07-d75844253ea1/kube-rbac-proxy/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.901061 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-2j9rb_42949778-256b-4809-b2ce-d633428e649c/manager/0.log" Oct 04 04:48:29 crc kubenswrapper[4726]: I1004 04:48:29.952830 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4v7k8_392e7503-f32e-4e8f-ac07-d75844253ea1/manager/0.log" Oct 04 04:48:30 crc kubenswrapper[4726]: I1004 04:48:30.040481 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-kn2pc_262abd5f-25d1-488f-b864-f9cc308a68f4/kube-rbac-proxy/0.log" Oct 04 04:48:30 crc kubenswrapper[4726]: I1004 04:48:30.087118 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-kn2pc_262abd5f-25d1-488f-b864-f9cc308a68f4/manager/0.log" Oct 04 04:48:34 crc kubenswrapper[4726]: I1004 04:48:34.187930 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:48:34 crc kubenswrapper[4726]: I1004 04:48:34.188553 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:48:45 crc kubenswrapper[4726]: I1004 04:48:45.643523 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qstzg_fa73eb1e-f78b-41f1-9be1-aeae979d66f9/control-plane-machine-set-operator/0.log" Oct 04 04:48:45 crc kubenswrapper[4726]: I1004 04:48:45.858140 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ffl97_7386c233-34f0-4eef-87d0-ac2b7b145897/kube-rbac-proxy/0.log" Oct 04 04:48:45 crc kubenswrapper[4726]: I1004 04:48:45.895758 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ffl97_7386c233-34f0-4eef-87d0-ac2b7b145897/machine-api-operator/0.log" Oct 04 04:48:57 crc kubenswrapper[4726]: I1004 04:48:57.565877 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vbkwl_53efe0a4-7f9a-4541-9677-12e0dcec292e/cert-manager-controller/0.log" Oct 04 04:48:58 crc kubenswrapper[4726]: I1004 04:48:58.010519 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-nkt2j_5e7b44b7-601a-4d34-a2f5-13662f3ef3cd/cert-manager-webhook/0.log" Oct 04 04:48:58 crc kubenswrapper[4726]: I1004 04:48:58.139155 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hkk79_3628e5db-697e-495c-815f-c08b5ccd03da/cert-manager-cainjector/0.log" Oct 04 04:49:04 crc kubenswrapper[4726]: I1004 04:49:04.188429 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:49:04 crc kubenswrapper[4726]: I1004 04:49:04.189223 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:49:09 crc kubenswrapper[4726]: I1004 04:49:09.650940 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-2tqv5_0afe634c-ed15-4abf-998a-5929ab4c94c3/nmstate-console-plugin/0.log" Oct 04 04:49:09 crc kubenswrapper[4726]: I1004 04:49:09.834476 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jx5pc_b0606264-b345-49e2-9965-806462686e84/nmstate-handler/0.log" Oct 04 04:49:09 crc kubenswrapper[4726]: I1004 04:49:09.904297 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8n4qb_10fdef1b-c247-4b30-82b9-cde45b292b17/kube-rbac-proxy/0.log" Oct 04 04:49:09 crc kubenswrapper[4726]: I1004 04:49:09.916038 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8n4qb_10fdef1b-c247-4b30-82b9-cde45b292b17/nmstate-metrics/0.log" Oct 04 04:49:10 crc kubenswrapper[4726]: I1004 04:49:10.038069 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-s4p69_a5ee8c2e-8f69-4301-8aab-c5af01402c8a/nmstate-operator/0.log" Oct 04 04:49:10 crc kubenswrapper[4726]: I1004 04:49:10.088533 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-mxq57_b2974204-71e0-4cd1-859b-0aef97909ee9/nmstate-webhook/0.log" Oct 04 04:49:22 crc kubenswrapper[4726]: I1004 04:49:22.953271 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tv4tz_eca76b37-13e9-4101-8486-7d009410a7bb/kube-rbac-proxy/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.018202 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tv4tz_eca76b37-13e9-4101-8486-7d009410a7bb/controller/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.186869 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.189383 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-76nkf_555989d8-4bc6-4750-a532-f71f09933d71/frr-k8s-webhook-server/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.405782 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.421762 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.422140 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.428049 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.634065 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.674609 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.674937 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.688925 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.841007 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.904502 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/controller/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.912711 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:49:23 crc kubenswrapper[4726]: I1004 04:49:23.932809 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.089991 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/frr-metrics/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.098660 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/kube-rbac-proxy/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.190459 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/kube-rbac-proxy-frr/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.340081 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/reloader/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.437331 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-655cdc4dc4-v2hjh_898913f8-1833-4c8f-8764-32c656d6caab/manager/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.542630 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66d48bf4d6-nkg69_898ba0b6-6440-45be-8067-6719ead41cba/webhook-server/0.log" Oct 04 04:49:24 crc kubenswrapper[4726]: I1004 04:49:24.739734 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89wr2_ca2e3d23-2c1d-4b4a-85d6-c280ce107167/kube-rbac-proxy/0.log" Oct 04 04:49:25 crc kubenswrapper[4726]: I1004 04:49:25.342315 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89wr2_ca2e3d23-2c1d-4b4a-85d6-c280ce107167/speaker/0.log" Oct 04 04:49:25 crc kubenswrapper[4726]: I1004 04:49:25.503431 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/frr/0.log" Oct 04 04:49:34 crc kubenswrapper[4726]: I1004 04:49:34.188654 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:49:34 crc kubenswrapper[4726]: I1004 04:49:34.189320 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:49:34 crc kubenswrapper[4726]: I1004 04:49:34.189377 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:49:34 crc kubenswrapper[4726]: I1004 04:49:34.190263 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:49:34 crc kubenswrapper[4726]: I1004 04:49:34.190327 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882" gracePeriod=600 Oct 04 04:49:35 crc kubenswrapper[4726]: I1004 04:49:35.273625 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882" exitCode=0 Oct 04 04:49:35 crc kubenswrapper[4726]: I1004 04:49:35.274288 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882"} Oct 04 04:49:35 crc kubenswrapper[4726]: I1004 04:49:35.274324 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec"} Oct 04 04:49:35 crc kubenswrapper[4726]: I1004 04:49:35.274344 4726 scope.go:117] "RemoveContainer" containerID="da94df0fba3c864a202d4fa07da1230794a9306f28d820fe4fe711eb4f3fc911" Oct 04 04:49:36 crc kubenswrapper[4726]: I1004 04:49:36.847368 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.037444 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.074887 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.104838 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.264692 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.265903 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/extract/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.267419 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.415498 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.593448 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.650000 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.691710 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.884213 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:49:37 crc kubenswrapper[4726]: I1004 04:49:37.886872 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.101530 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.349299 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.391728 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.405893 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.455819 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/registry-server/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.612955 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.621342 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:49:38 crc kubenswrapper[4726]: I1004 04:49:38.866185 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.405655 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.405729 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.510423 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.623994 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.673334 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.723067 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/extract/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.729836 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/registry-server/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.873352 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mqgpt_f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8/marketplace-operator/0.log" Oct 04 04:49:39 crc kubenswrapper[4726]: I1004 04:49:39.941608 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.146764 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.164775 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.181248 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.312271 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.330831 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.357865 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.501378 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/registry-server/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.587840 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.636966 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.655650 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.800735 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:49:40 crc kubenswrapper[4726]: I1004 04:49:40.807692 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:49:41 crc kubenswrapper[4726]: I1004 04:49:41.498244 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/registry-server/0.log" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.157318 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:07 crc kubenswrapper[4726]: E1004 04:51:07.158526 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" containerName="container-00" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.158551 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" containerName="container-00" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.158978 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="347b3f34-2be9-4bcf-803d-a67fc3cd7e3c" containerName="container-00" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.161610 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.169723 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.184824 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.185142 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.185315 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4bt5\" (UniqueName: \"kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.287347 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4bt5\" (UniqueName: \"kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.287493 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.287549 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.288231 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.288317 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.306424 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4bt5\" (UniqueName: \"kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5\") pod \"redhat-marketplace-9rvnr\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.480609 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:07 crc kubenswrapper[4726]: I1004 04:51:07.955901 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:08 crc kubenswrapper[4726]: I1004 04:51:08.094254 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerStarted","Data":"d02d2c669ea9b1ac99db1320e26c04249c80454cc0abf5d7b618ffb600694618"} Oct 04 04:51:09 crc kubenswrapper[4726]: I1004 04:51:09.106958 4726 generic.go:334] "Generic (PLEG): container finished" podID="92f2bd76-93a8-4354-993d-111b0c622b74" containerID="3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99" exitCode=0 Oct 04 04:51:09 crc kubenswrapper[4726]: I1004 04:51:09.107120 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerDied","Data":"3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99"} Oct 04 04:51:09 crc kubenswrapper[4726]: I1004 04:51:09.109511 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:51:10 crc kubenswrapper[4726]: I1004 04:51:10.121928 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerStarted","Data":"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97"} Oct 04 04:51:11 crc kubenswrapper[4726]: I1004 04:51:11.142485 4726 generic.go:334] "Generic (PLEG): container finished" podID="92f2bd76-93a8-4354-993d-111b0c622b74" containerID="681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97" exitCode=0 Oct 04 04:51:11 crc kubenswrapper[4726]: I1004 04:51:11.142639 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerDied","Data":"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97"} Oct 04 04:51:13 crc kubenswrapper[4726]: I1004 04:51:13.166325 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerStarted","Data":"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454"} Oct 04 04:51:14 crc kubenswrapper[4726]: I1004 04:51:14.198722 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9rvnr" podStartSLOduration=3.530414076 podStartE2EDuration="7.198704793s" podCreationTimestamp="2025-10-04 04:51:07 +0000 UTC" firstStartedPulling="2025-10-04 04:51:09.109328048 +0000 UTC m=+4243.283951261" lastFinishedPulling="2025-10-04 04:51:12.777618725 +0000 UTC m=+4246.952241978" observedRunningTime="2025-10-04 04:51:14.191166355 +0000 UTC m=+4248.365789578" watchObservedRunningTime="2025-10-04 04:51:14.198704793 +0000 UTC m=+4248.373328006" Oct 04 04:51:17 crc kubenswrapper[4726]: I1004 04:51:17.481769 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:17 crc kubenswrapper[4726]: I1004 04:51:17.482571 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:17 crc kubenswrapper[4726]: I1004 04:51:17.567037 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:18 crc kubenswrapper[4726]: I1004 04:51:18.271667 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:18 crc kubenswrapper[4726]: I1004 04:51:18.326789 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.233289 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9rvnr" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="registry-server" containerID="cri-o://9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454" gracePeriod=2 Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.678316 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.849179 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content\") pod \"92f2bd76-93a8-4354-993d-111b0c622b74\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.849650 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities\") pod \"92f2bd76-93a8-4354-993d-111b0c622b74\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.849905 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4bt5\" (UniqueName: \"kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5\") pod \"92f2bd76-93a8-4354-993d-111b0c622b74\" (UID: \"92f2bd76-93a8-4354-993d-111b0c622b74\") " Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.850614 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities" (OuterVolumeSpecName: "utilities") pod "92f2bd76-93a8-4354-993d-111b0c622b74" (UID: "92f2bd76-93a8-4354-993d-111b0c622b74"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.868181 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5" (OuterVolumeSpecName: "kube-api-access-r4bt5") pod "92f2bd76-93a8-4354-993d-111b0c622b74" (UID: "92f2bd76-93a8-4354-993d-111b0c622b74"). InnerVolumeSpecName "kube-api-access-r4bt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.922961 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92f2bd76-93a8-4354-993d-111b0c622b74" (UID: "92f2bd76-93a8-4354-993d-111b0c622b74"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.952415 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4bt5\" (UniqueName: \"kubernetes.io/projected/92f2bd76-93a8-4354-993d-111b0c622b74-kube-api-access-r4bt5\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.952636 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:20 crc kubenswrapper[4726]: I1004 04:51:20.952707 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f2bd76-93a8-4354-993d-111b0c622b74-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.250291 4726 generic.go:334] "Generic (PLEG): container finished" podID="92f2bd76-93a8-4354-993d-111b0c622b74" containerID="9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454" exitCode=0 Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.250351 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerDied","Data":"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454"} Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.250387 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9rvnr" event={"ID":"92f2bd76-93a8-4354-993d-111b0c622b74","Type":"ContainerDied","Data":"d02d2c669ea9b1ac99db1320e26c04249c80454cc0abf5d7b618ffb600694618"} Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.250408 4726 scope.go:117] "RemoveContainer" containerID="9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.250606 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9rvnr" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.279403 4726 scope.go:117] "RemoveContainer" containerID="681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.286302 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.294737 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9rvnr"] Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.375785 4726 scope.go:117] "RemoveContainer" containerID="3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.393871 4726 scope.go:117] "RemoveContainer" containerID="9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454" Oct 04 04:51:21 crc kubenswrapper[4726]: E1004 04:51:21.394499 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454\": container with ID starting with 9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454 not found: ID does not exist" containerID="9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.394552 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454"} err="failed to get container status \"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454\": rpc error: code = NotFound desc = could not find container \"9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454\": container with ID starting with 9737369087cc1f31212719882294f7aa1340f131375378704d24fa39f8a06454 not found: ID does not exist" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.394618 4726 scope.go:117] "RemoveContainer" containerID="681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97" Oct 04 04:51:21 crc kubenswrapper[4726]: E1004 04:51:21.395304 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97\": container with ID starting with 681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97 not found: ID does not exist" containerID="681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.395384 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97"} err="failed to get container status \"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97\": rpc error: code = NotFound desc = could not find container \"681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97\": container with ID starting with 681fbddd6bc99336a1cf0908b1e522f8b962ec65bd064023355b2eacde56fd97 not found: ID does not exist" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.395413 4726 scope.go:117] "RemoveContainer" containerID="3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99" Oct 04 04:51:21 crc kubenswrapper[4726]: E1004 04:51:21.395685 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99\": container with ID starting with 3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99 not found: ID does not exist" containerID="3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99" Oct 04 04:51:21 crc kubenswrapper[4726]: I1004 04:51:21.395711 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99"} err="failed to get container status \"3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99\": rpc error: code = NotFound desc = could not find container \"3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99\": container with ID starting with 3389d847d696506cbc57613b2eafb9af169d136bd7ed2883aa7c104254591b99 not found: ID does not exist" Oct 04 04:51:22 crc kubenswrapper[4726]: I1004 04:51:22.524901 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" path="/var/lib/kubelet/pods/92f2bd76-93a8-4354-993d-111b0c622b74/volumes" Oct 04 04:51:34 crc kubenswrapper[4726]: I1004 04:51:34.187924 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:51:34 crc kubenswrapper[4726]: I1004 04:51:34.188634 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:51:46 crc kubenswrapper[4726]: I1004 04:51:46.495080 4726 generic.go:334] "Generic (PLEG): container finished" podID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerID="9075792a1a4d2f84a04b7de3665ec39cd74712ee6d7fbe5b0ff6c1f09659fe8b" exitCode=0 Oct 04 04:51:46 crc kubenswrapper[4726]: I1004 04:51:46.495231 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" event={"ID":"566782cd-404b-4bb9-9b0f-7d3892e15c0e","Type":"ContainerDied","Data":"9075792a1a4d2f84a04b7de3665ec39cd74712ee6d7fbe5b0ff6c1f09659fe8b"} Oct 04 04:51:46 crc kubenswrapper[4726]: I1004 04:51:46.496357 4726 scope.go:117] "RemoveContainer" containerID="9075792a1a4d2f84a04b7de3665ec39cd74712ee6d7fbe5b0ff6c1f09659fe8b" Oct 04 04:51:47 crc kubenswrapper[4726]: I1004 04:51:47.227463 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s7tm6_must-gather-7fvw9_566782cd-404b-4bb9-9b0f-7d3892e15c0e/gather/0.log" Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.392021 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-s7tm6/must-gather-7fvw9"] Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.392825 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="copy" containerID="cri-o://8558bae34980123ba2f6c197b667a27b5a1ec7e1110ef3dd132f9e9cc1ff2408" gracePeriod=2 Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.402386 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-s7tm6/must-gather-7fvw9"] Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.582418 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s7tm6_must-gather-7fvw9_566782cd-404b-4bb9-9b0f-7d3892e15c0e/copy/0.log" Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.582804 4726 generic.go:334] "Generic (PLEG): container finished" podID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerID="8558bae34980123ba2f6c197b667a27b5a1ec7e1110ef3dd132f9e9cc1ff2408" exitCode=143 Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.980052 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s7tm6_must-gather-7fvw9_566782cd-404b-4bb9-9b0f-7d3892e15c0e/copy/0.log" Oct 04 04:51:55 crc kubenswrapper[4726]: I1004 04:51:55.980795 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.172981 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output\") pod \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.173047 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cw4d\" (UniqueName: \"kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d\") pod \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\" (UID: \"566782cd-404b-4bb9-9b0f-7d3892e15c0e\") " Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.183389 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d" (OuterVolumeSpecName: "kube-api-access-8cw4d") pod "566782cd-404b-4bb9-9b0f-7d3892e15c0e" (UID: "566782cd-404b-4bb9-9b0f-7d3892e15c0e"). InnerVolumeSpecName "kube-api-access-8cw4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.279658 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cw4d\" (UniqueName: \"kubernetes.io/projected/566782cd-404b-4bb9-9b0f-7d3892e15c0e-kube-api-access-8cw4d\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.349005 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "566782cd-404b-4bb9-9b0f-7d3892e15c0e" (UID: "566782cd-404b-4bb9-9b0f-7d3892e15c0e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.381136 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/566782cd-404b-4bb9-9b0f-7d3892e15c0e-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.513715 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" path="/var/lib/kubelet/pods/566782cd-404b-4bb9-9b0f-7d3892e15c0e/volumes" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.593815 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-s7tm6_must-gather-7fvw9_566782cd-404b-4bb9-9b0f-7d3892e15c0e/copy/0.log" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.594211 4726 scope.go:117] "RemoveContainer" containerID="8558bae34980123ba2f6c197b667a27b5a1ec7e1110ef3dd132f9e9cc1ff2408" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.594350 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-s7tm6/must-gather-7fvw9" Oct 04 04:51:56 crc kubenswrapper[4726]: I1004 04:51:56.622340 4726 scope.go:117] "RemoveContainer" containerID="9075792a1a4d2f84a04b7de3665ec39cd74712ee6d7fbe5b0ff6c1f09659fe8b" Oct 04 04:51:56 crc kubenswrapper[4726]: E1004 04:51:56.696915 4726 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566782cd_404b_4bb9_9b0f_7d3892e15c0e.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:52:04 crc kubenswrapper[4726]: I1004 04:52:04.188361 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:52:04 crc kubenswrapper[4726]: I1004 04:52:04.189031 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.089140 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:09 crc kubenswrapper[4726]: E1004 04:52:09.090396 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="registry-server" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090417 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="registry-server" Oct 04 04:52:09 crc kubenswrapper[4726]: E1004 04:52:09.090439 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="extract-content" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090448 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="extract-content" Oct 04 04:52:09 crc kubenswrapper[4726]: E1004 04:52:09.090469 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="gather" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090479 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="gather" Oct 04 04:52:09 crc kubenswrapper[4726]: E1004 04:52:09.090489 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="copy" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090496 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="copy" Oct 04 04:52:09 crc kubenswrapper[4726]: E1004 04:52:09.090518 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="extract-utilities" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090526 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="extract-utilities" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090761 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="gather" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090780 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f2bd76-93a8-4354-993d-111b0c622b74" containerName="registry-server" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.090796 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="566782cd-404b-4bb9-9b0f-7d3892e15c0e" containerName="copy" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.092189 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.106896 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.205511 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.208031 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.209316 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc8lq\" (UniqueName: \"kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.312082 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc8lq\" (UniqueName: \"kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.312424 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.312709 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.312981 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.313376 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.335391 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc8lq\" (UniqueName: \"kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq\") pod \"certified-operators-5q584\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.422817 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:09 crc kubenswrapper[4726]: I1004 04:52:09.940051 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:09 crc kubenswrapper[4726]: W1004 04:52:09.945677 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda214f563_67d4_46d6_bad3_1bb66702e6a2.slice/crio-3c019d9dfeacc8b1fb37712081dd526232e3fcfe01c97733f80e9118f341b0bb WatchSource:0}: Error finding container 3c019d9dfeacc8b1fb37712081dd526232e3fcfe01c97733f80e9118f341b0bb: Status 404 returned error can't find the container with id 3c019d9dfeacc8b1fb37712081dd526232e3fcfe01c97733f80e9118f341b0bb Oct 04 04:52:10 crc kubenswrapper[4726]: I1004 04:52:10.713135 4726 generic.go:334] "Generic (PLEG): container finished" podID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerID="a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511" exitCode=0 Oct 04 04:52:10 crc kubenswrapper[4726]: I1004 04:52:10.713239 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerDied","Data":"a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511"} Oct 04 04:52:10 crc kubenswrapper[4726]: I1004 04:52:10.713637 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerStarted","Data":"3c019d9dfeacc8b1fb37712081dd526232e3fcfe01c97733f80e9118f341b0bb"} Oct 04 04:52:12 crc kubenswrapper[4726]: I1004 04:52:12.735839 4726 generic.go:334] "Generic (PLEG): container finished" podID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerID="41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509" exitCode=0 Oct 04 04:52:12 crc kubenswrapper[4726]: I1004 04:52:12.735885 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerDied","Data":"41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509"} Oct 04 04:52:14 crc kubenswrapper[4726]: I1004 04:52:14.756011 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerStarted","Data":"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680"} Oct 04 04:52:14 crc kubenswrapper[4726]: I1004 04:52:14.777026 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5q584" podStartSLOduration=3.156493987 podStartE2EDuration="5.777007365s" podCreationTimestamp="2025-10-04 04:52:09 +0000 UTC" firstStartedPulling="2025-10-04 04:52:10.71518568 +0000 UTC m=+4304.889808893" lastFinishedPulling="2025-10-04 04:52:13.335699058 +0000 UTC m=+4307.510322271" observedRunningTime="2025-10-04 04:52:14.772835436 +0000 UTC m=+4308.947458649" watchObservedRunningTime="2025-10-04 04:52:14.777007365 +0000 UTC m=+4308.951630578" Oct 04 04:52:19 crc kubenswrapper[4726]: I1004 04:52:19.423959 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:19 crc kubenswrapper[4726]: I1004 04:52:19.424905 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:19 crc kubenswrapper[4726]: I1004 04:52:19.469651 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:19 crc kubenswrapper[4726]: I1004 04:52:19.852955 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:19 crc kubenswrapper[4726]: I1004 04:52:19.898406 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:21 crc kubenswrapper[4726]: I1004 04:52:21.819492 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5q584" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="registry-server" containerID="cri-o://d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680" gracePeriod=2 Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.279238 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.409270 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities\") pod \"a214f563-67d4-46d6-bad3-1bb66702e6a2\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.409354 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc8lq\" (UniqueName: \"kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq\") pod \"a214f563-67d4-46d6-bad3-1bb66702e6a2\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.409391 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content\") pod \"a214f563-67d4-46d6-bad3-1bb66702e6a2\" (UID: \"a214f563-67d4-46d6-bad3-1bb66702e6a2\") " Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.410383 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities" (OuterVolumeSpecName: "utilities") pod "a214f563-67d4-46d6-bad3-1bb66702e6a2" (UID: "a214f563-67d4-46d6-bad3-1bb66702e6a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.417699 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq" (OuterVolumeSpecName: "kube-api-access-wc8lq") pod "a214f563-67d4-46d6-bad3-1bb66702e6a2" (UID: "a214f563-67d4-46d6-bad3-1bb66702e6a2"). InnerVolumeSpecName "kube-api-access-wc8lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.466695 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a214f563-67d4-46d6-bad3-1bb66702e6a2" (UID: "a214f563-67d4-46d6-bad3-1bb66702e6a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.511420 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.511450 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc8lq\" (UniqueName: \"kubernetes.io/projected/a214f563-67d4-46d6-bad3-1bb66702e6a2-kube-api-access-wc8lq\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.511459 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a214f563-67d4-46d6-bad3-1bb66702e6a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.833547 4726 generic.go:334] "Generic (PLEG): container finished" podID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerID="d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680" exitCode=0 Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.833592 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerDied","Data":"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680"} Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.833614 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q584" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.833633 4726 scope.go:117] "RemoveContainer" containerID="d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.833622 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q584" event={"ID":"a214f563-67d4-46d6-bad3-1bb66702e6a2","Type":"ContainerDied","Data":"3c019d9dfeacc8b1fb37712081dd526232e3fcfe01c97733f80e9118f341b0bb"} Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.865508 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.868800 4726 scope.go:117] "RemoveContainer" containerID="41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.874931 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5q584"] Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.887439 4726 scope.go:117] "RemoveContainer" containerID="a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.931010 4726 scope.go:117] "RemoveContainer" containerID="d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680" Oct 04 04:52:22 crc kubenswrapper[4726]: E1004 04:52:22.931620 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680\": container with ID starting with d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680 not found: ID does not exist" containerID="d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.931673 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680"} err="failed to get container status \"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680\": rpc error: code = NotFound desc = could not find container \"d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680\": container with ID starting with d31fcbd8d6f6aba35d901cb318847f3f3d1abc31b0ee900ac9e1129165120680 not found: ID does not exist" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.931701 4726 scope.go:117] "RemoveContainer" containerID="41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509" Oct 04 04:52:22 crc kubenswrapper[4726]: E1004 04:52:22.932145 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509\": container with ID starting with 41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509 not found: ID does not exist" containerID="41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.932199 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509"} err="failed to get container status \"41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509\": rpc error: code = NotFound desc = could not find container \"41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509\": container with ID starting with 41f509721f6f788be45d47f0f2c25a5a1a3bacf0bfae016688a191e34cb16509 not found: ID does not exist" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.932233 4726 scope.go:117] "RemoveContainer" containerID="a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511" Oct 04 04:52:22 crc kubenswrapper[4726]: E1004 04:52:22.932755 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511\": container with ID starting with a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511 not found: ID does not exist" containerID="a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511" Oct 04 04:52:22 crc kubenswrapper[4726]: I1004 04:52:22.932785 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511"} err="failed to get container status \"a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511\": rpc error: code = NotFound desc = could not find container \"a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511\": container with ID starting with a58d1961b82b337caf7c246cf87a4fa91763730ca3d59cc8b799f12332eac511 not found: ID does not exist" Oct 04 04:52:24 crc kubenswrapper[4726]: I1004 04:52:24.513942 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" path="/var/lib/kubelet/pods/a214f563-67d4-46d6-bad3-1bb66702e6a2/volumes" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.188040 4726 patch_prober.go:28] interesting pod/machine-config-daemon-9777l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.188818 4726 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.188892 4726 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9777l" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.190004 4726 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec"} pod="openshift-machine-config-operator/machine-config-daemon-9777l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.190175 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerName="machine-config-daemon" containerID="cri-o://1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" gracePeriod=600 Oct 04 04:52:34 crc kubenswrapper[4726]: E1004 04:52:34.322392 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.955294 4726 generic.go:334] "Generic (PLEG): container finished" podID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" exitCode=0 Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.955337 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerDied","Data":"1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec"} Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.955374 4726 scope.go:117] "RemoveContainer" containerID="aca27ff3e5bd9e906469b91ae5ccf09c3a7506936be0e5046ff63d2261102882" Oct 04 04:52:34 crc kubenswrapper[4726]: I1004 04:52:34.956060 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:52:34 crc kubenswrapper[4726]: E1004 04:52:34.956354 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.575342 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lk69/must-gather-lwbct"] Oct 04 04:52:35 crc kubenswrapper[4726]: E1004 04:52:35.576026 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="extract-content" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.576038 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="extract-content" Oct 04 04:52:35 crc kubenswrapper[4726]: E1004 04:52:35.576052 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="registry-server" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.576058 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="registry-server" Oct 04 04:52:35 crc kubenswrapper[4726]: E1004 04:52:35.576076 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="extract-utilities" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.576082 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="extract-utilities" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.576283 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="a214f563-67d4-46d6-bad3-1bb66702e6a2" containerName="registry-server" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.577256 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.585618 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7lk69"/"openshift-service-ca.crt" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.585877 4726 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7lk69"/"kube-root-ca.crt" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.586059 4726 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7lk69"/"default-dockercfg-zqth9" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.596932 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7lk69/must-gather-lwbct"] Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.674512 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnrf6\" (UniqueName: \"kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.674783 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.776641 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnrf6\" (UniqueName: \"kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.776696 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.777093 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.803446 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnrf6\" (UniqueName: \"kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6\") pod \"must-gather-lwbct\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:35 crc kubenswrapper[4726]: I1004 04:52:35.926248 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:52:36 crc kubenswrapper[4726]: I1004 04:52:36.491119 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7lk69/must-gather-lwbct"] Oct 04 04:52:37 crc kubenswrapper[4726]: I1004 04:52:37.006631 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/must-gather-lwbct" event={"ID":"97accf81-66a4-49ff-8228-0505dc1bfd94","Type":"ContainerStarted","Data":"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1"} Oct 04 04:52:37 crc kubenswrapper[4726]: I1004 04:52:37.008659 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/must-gather-lwbct" event={"ID":"97accf81-66a4-49ff-8228-0505dc1bfd94","Type":"ContainerStarted","Data":"cfe73badab3ce58b844b0b8449dcd7252288f96cd7f7d3ce4e3a4a205b042dfa"} Oct 04 04:52:38 crc kubenswrapper[4726]: I1004 04:52:38.025699 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/must-gather-lwbct" event={"ID":"97accf81-66a4-49ff-8228-0505dc1bfd94","Type":"ContainerStarted","Data":"86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5"} Oct 04 04:52:38 crc kubenswrapper[4726]: I1004 04:52:38.057535 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7lk69/must-gather-lwbct" podStartSLOduration=3.057504212 podStartE2EDuration="3.057504212s" podCreationTimestamp="2025-10-04 04:52:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:38.04253404 +0000 UTC m=+4332.217157273" watchObservedRunningTime="2025-10-04 04:52:38.057504212 +0000 UTC m=+4332.232127465" Oct 04 04:52:40 crc kubenswrapper[4726]: I1004 04:52:40.853068 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lk69/crc-debug-tqq28"] Oct 04 04:52:40 crc kubenswrapper[4726]: I1004 04:52:40.854703 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.021875 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.022047 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcrlr\" (UniqueName: \"kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.123855 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcrlr\" (UniqueName: \"kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.123936 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.124079 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.143409 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcrlr\" (UniqueName: \"kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr\") pod \"crc-debug-tqq28\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: I1004 04:52:41.174419 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:52:41 crc kubenswrapper[4726]: W1004 04:52:41.209953 4726 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd49937fb_d320_4c40_8727_a0d224e719e2.slice/crio-ca75e1e57fc8ace009322bcc4e7d4b04f52843968e745d09401a856aee5889af WatchSource:0}: Error finding container ca75e1e57fc8ace009322bcc4e7d4b04f52843968e745d09401a856aee5889af: Status 404 returned error can't find the container with id ca75e1e57fc8ace009322bcc4e7d4b04f52843968e745d09401a856aee5889af Oct 04 04:52:42 crc kubenswrapper[4726]: I1004 04:52:42.060169 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-tqq28" event={"ID":"d49937fb-d320-4c40-8727-a0d224e719e2","Type":"ContainerStarted","Data":"ff253348e07af6b66ac29dd7f071f26b87427e8e3545f3008e1d11ce3bb5de36"} Oct 04 04:52:42 crc kubenswrapper[4726]: I1004 04:52:42.060782 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-tqq28" event={"ID":"d49937fb-d320-4c40-8727-a0d224e719e2","Type":"ContainerStarted","Data":"ca75e1e57fc8ace009322bcc4e7d4b04f52843968e745d09401a856aee5889af"} Oct 04 04:52:42 crc kubenswrapper[4726]: I1004 04:52:42.075204 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7lk69/crc-debug-tqq28" podStartSLOduration=2.07518035 podStartE2EDuration="2.07518035s" podCreationTimestamp="2025-10-04 04:52:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:42.074512803 +0000 UTC m=+4336.249136016" watchObservedRunningTime="2025-10-04 04:52:42.07518035 +0000 UTC m=+4336.249803553" Oct 04 04:52:48 crc kubenswrapper[4726]: I1004 04:52:48.504021 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:52:48 crc kubenswrapper[4726]: E1004 04:52:48.505411 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:52:53 crc kubenswrapper[4726]: I1004 04:52:53.871181 4726 scope.go:117] "RemoveContainer" containerID="54d4ab076a0f3510889bcc7e6485e84a2f4f77b6bef96b4ecf71fab7d00aecce" Oct 04 04:53:03 crc kubenswrapper[4726]: I1004 04:53:03.503130 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:53:03 crc kubenswrapper[4726]: E1004 04:53:03.503829 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:53:15 crc kubenswrapper[4726]: I1004 04:53:15.503379 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:53:15 crc kubenswrapper[4726]: E1004 04:53:15.504096 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:53:29 crc kubenswrapper[4726]: I1004 04:53:29.502798 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:53:29 crc kubenswrapper[4726]: E1004 04:53:29.503651 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:53:43 crc kubenswrapper[4726]: I1004 04:53:43.502652 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:53:43 crc kubenswrapper[4726]: E1004 04:53:43.503501 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.160633 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56bb546-6dhls_d74c5864-48ad-4376-b318-071c1925ad4c/barbican-api-log/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.187375 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-56bb546-6dhls_d74c5864-48ad-4376-b318-071c1925ad4c/barbican-api/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.353018 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fccfff78b-bs7hd_0f684dad-f165-4648-a5b0-e1e94b2cdc44/barbican-keystone-listener/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.409501 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5fccfff78b-bs7hd_0f684dad-f165-4648-a5b0-e1e94b2cdc44/barbican-keystone-listener-log/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.583722 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78dd78b9bc-pwdzf_be9d885c-08ac-43eb-a1a7-2829719d4187/barbican-worker/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.641502 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78dd78b9bc-pwdzf_be9d885c-08ac-43eb-a1a7-2829719d4187/barbican-worker-log/0.log" Oct 04 04:53:49 crc kubenswrapper[4726]: I1004 04:53:49.895781 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-hktvg_aa6f54c4-9900-41b2-a047-6dda8304ff99/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.057194 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/ceilometer-central-agent/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.092029 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/proxy-httpd/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.136675 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/ceilometer-notification-agent/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.284558 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_00a3dd9a-1341-4141-a0e6-8dc5b3e73565/sg-core/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.398125 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cc91ace6-aab5-4a5c-9549-131d3303f5d5/cinder-api/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.717972 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cc91ace6-aab5-4a5c-9549-131d3303f5d5/cinder-api-log/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.852830 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2105d235-5b6d-40ea-80d4-0f2033e0bf5d/cinder-scheduler/0.log" Oct 04 04:53:50 crc kubenswrapper[4726]: I1004 04:53:50.945137 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2105d235-5b6d-40ea-80d4-0f2033e0bf5d/probe/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.052259 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jwwh7_ebf88812-a311-4c5a-8698-d50f551446f2/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.172274 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-l8c7q_13859a35-3c01-4314-a416-0d3d04828cdc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.249868 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-rtkz5_dec26676-436d-4d9a-8337-0c9ac86c6168/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.410415 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/init/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.543998 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/init/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.598939 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-snnnq_0b74ac9e-0f00-4a57-a05c-a5a155fd51f9/dnsmasq-dns/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.627925 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6gwh5_4a4f5e93-bd73-42f9-8282-39e559ad08a7/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.799698 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bae4a133-24a9-4ba4-8050-5e1a15bc8ed3/glance-httpd/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.801504 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bae4a133-24a9-4ba4-8050-5e1a15bc8ed3/glance-log/0.log" Oct 04 04:53:51 crc kubenswrapper[4726]: I1004 04:53:51.961918 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_636ad703-b66b-42f6-aca7-00900d0fad9b/glance-httpd/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.009247 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_636ad703-b66b-42f6-aca7-00900d0fad9b/glance-log/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.188244 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7786888bd-6khpf_888736e3-9e3d-402b-9ea3-79a0ba740fe7/horizon/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.306952 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nzrft_70435c0c-e859-49bd-9e2e-624f66c3be54/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.571196 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-v85jc_1faabfae-0c50-425d-af5b-529a69ee6791/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.673771 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7786888bd-6khpf_888736e3-9e3d-402b-9ea3-79a0ba740fe7/horizon-log/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.913579 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325841-fnfsx_db67cb41-86ca-41f3-b012-1bfc33e84a37/keystone-cron/0.log" Oct 04 04:53:52 crc kubenswrapper[4726]: I1004 04:53:52.922561 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-79b54b49b-mx2kb_cb8197e3-5f74-443a-bf22-78a28365cfaf/keystone-api/0.log" Oct 04 04:53:53 crc kubenswrapper[4726]: I1004 04:53:53.026720 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c4db3f45-a9c5-4092-a236-f3c345e2303d/kube-state-metrics/0.log" Oct 04 04:53:53 crc kubenswrapper[4726]: I1004 04:53:53.131099 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-5ps4d_836b7c28-6c4b-44c1-9f6f-ff52d2d7b107/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:53 crc kubenswrapper[4726]: I1004 04:53:53.549684 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5bd64b8469-btvtm_d97c6241-e072-4012-a61e-c5e855407e99/neutron-api/0.log" Oct 04 04:53:53 crc kubenswrapper[4726]: I1004 04:53:53.565393 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5bd64b8469-btvtm_d97c6241-e072-4012-a61e-c5e855407e99/neutron-httpd/0.log" Oct 04 04:53:53 crc kubenswrapper[4726]: I1004 04:53:53.763623 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fpkwz_cbf08126-f813-40fe-b451-71cab2c2fd8a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:54 crc kubenswrapper[4726]: I1004 04:53:54.464951 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5c95d78d-1f21-48f0-903e-713cde259066/nova-api-log/0.log" Oct 04 04:53:54 crc kubenswrapper[4726]: I1004 04:53:54.657893 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_76642298-36b4-4bde-ba49-3b0a2e56494d/nova-cell0-conductor-conductor/0.log" Oct 04 04:53:54 crc kubenswrapper[4726]: I1004 04:53:54.764010 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5c95d78d-1f21-48f0-903e-713cde259066/nova-api-api/0.log" Oct 04 04:53:55 crc kubenswrapper[4726]: I1004 04:53:55.072845 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_11f9ad5d-1851-40d1-b8f9-cfd476b498ce/nova-cell1-conductor-conductor/0.log" Oct 04 04:53:55 crc kubenswrapper[4726]: I1004 04:53:55.148701 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_df4fee1d-47f1-48e0-98bd-f4db94f5a543/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 04:53:55 crc kubenswrapper[4726]: I1004 04:53:55.411726 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-j6dh8_cb83849c-9bbd-479b-9a49-e99ea5247ab1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:55 crc kubenswrapper[4726]: I1004 04:53:55.485831 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f13faed-0be2-4916-80ab-f9d7c7d196a3/nova-metadata-log/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.007092 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2a84708a-3dbc-477f-b151-d68b83b4617b/nova-scheduler-scheduler/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.185789 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/mysql-bootstrap/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.393592 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/mysql-bootstrap/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.405926 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ed4e96c4-5868-4fd7-970b-ca7c74b56b1a/galera/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.632224 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/mysql-bootstrap/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.891993 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/mysql-bootstrap/0.log" Oct 04 04:53:56 crc kubenswrapper[4726]: I1004 04:53:56.974292 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d148624b-4c10-4a20-b428-902273d70ac0/galera/0.log" Oct 04 04:53:57 crc kubenswrapper[4726]: I1004 04:53:57.190193 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f592dceb-e227-48ea-a6d6-0b9513ec5b46/openstackclient/0.log" Oct 04 04:53:57 crc kubenswrapper[4726]: I1004 04:53:57.395584 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1f13faed-0be2-4916-80ab-f9d7c7d196a3/nova-metadata-metadata/0.log" Oct 04 04:53:57 crc kubenswrapper[4726]: I1004 04:53:57.502627 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:53:57 crc kubenswrapper[4726]: E1004 04:53:57.502889 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:53:57 crc kubenswrapper[4726]: I1004 04:53:57.954388 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5kwdw_8675d591-017c-4479-bcb2-dbb70c2f9f93/ovn-controller/0.log" Oct 04 04:53:57 crc kubenswrapper[4726]: I1004 04:53:57.980929 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-r526s_507a6d8a-447d-436c-918d-7e2f447705bf/openstack-network-exporter/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.227674 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server-init/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.422347 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovs-vswitchd/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.424140 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server-init/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.484142 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7724w_178e673b-646c-44a4-b368-29d3fabce0c4/ovsdb-server/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.654945 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-jfxbp_e1087fbc-edc3-47d2-8229-fb1654ce31eb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.879319 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f3ae4630-e488-43db-a57a-762c75de75aa/ovn-northd/0.log" Oct 04 04:53:58 crc kubenswrapper[4726]: I1004 04:53:58.894505 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_f3ae4630-e488-43db-a57a-762c75de75aa/openstack-network-exporter/0.log" Oct 04 04:53:59 crc kubenswrapper[4726]: I1004 04:53:59.117090 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_164dc913-a191-4c88-ad58-e2a08341d0a3/openstack-network-exporter/0.log" Oct 04 04:53:59 crc kubenswrapper[4726]: I1004 04:53:59.146405 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_164dc913-a191-4c88-ad58-e2a08341d0a3/ovsdbserver-nb/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.003159 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d298c1d3-002d-4847-bddc-63df803a9029/openstack-network-exporter/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.070486 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d298c1d3-002d-4847-bddc-63df803a9029/ovsdbserver-sb/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.263964 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f697b85fb-mcpgk_c10188ee-0fc9-4985-a21c-88fbc85e1e3c/placement-api/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.354732 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f697b85fb-mcpgk_c10188ee-0fc9-4985-a21c-88fbc85e1e3c/placement-log/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.469949 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/setup-container/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.681120 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/rabbitmq/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.682433 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_179e2c3b-cbc2-4417-9228-1b5eea2a1ced/setup-container/0.log" Oct 04 04:54:00 crc kubenswrapper[4726]: I1004 04:54:00.873374 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/setup-container/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.058322 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/setup-container/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.072100 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4ede7a9f-b13d-4fa4-b5b4-9a5f78826351/rabbitmq/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.252994 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-w8969_0c4effe5-c8f2-4f82-b5b0-3052858642f1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.420072 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-6mqds_2aa7ced4-f233-4e8f-b1da-b3ff5933e9ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.517153 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5gnj2_db0467b4-3329-4278-b094-02b70e240f01/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.710414 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-786b8_9465de2d-b388-410c-90da-31beba8b5dbe/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:01 crc kubenswrapper[4726]: I1004 04:54:01.899801 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mjmvr_e0ad6b26-b887-46ea-9402-1e9bf3a77de9/ssh-known-hosts-edpm-deployment/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.169802 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bdb856b6c-f5xfb_503e5303-a481-4184-9bb0-2369123d5267/proxy-server/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.248317 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5bdb856b6c-f5xfb_503e5303-a481-4184-9bb0-2369123d5267/proxy-httpd/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.410223 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qxprg_63b2f4dc-088d-47ba-aad0-e0de06d93828/swift-ring-rebalance/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.449800 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-auditor/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.623378 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-reaper/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.679434 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-replicator/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.696455 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/account-server/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.819630 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-auditor/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.924236 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-replicator/0.log" Oct 04 04:54:02 crc kubenswrapper[4726]: I1004 04:54:02.930600 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-server/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.066355 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/container-updater/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.127499 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-expirer/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.206810 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-auditor/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.341214 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-server/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.362943 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-replicator/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.446235 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/object-updater/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.558322 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/rsync/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.650520 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c4528cfc-3f8b-4850-aeaa-eabe209382bd/swift-recon-cron/0.log" Oct 04 04:54:03 crc kubenswrapper[4726]: I1004 04:54:03.865471 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wclx9_e9a93781-fcc5-4323-bb02-934800e8de8f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:04 crc kubenswrapper[4726]: I1004 04:54:04.008414 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_de0aaf35-392c-4705-b554-1d4768c5cff1/tempest-tests-tempest-tests-runner/0.log" Oct 04 04:54:04 crc kubenswrapper[4726]: I1004 04:54:04.245840 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c20315a1-0428-4211-b771-49a5504712dd/test-operator-logs-container/0.log" Oct 04 04:54:04 crc kubenswrapper[4726]: I1004 04:54:04.384460 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5wb7w_affede2f-dbd9-44c4-ae15-af1fb96ce9f6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 04:54:09 crc kubenswrapper[4726]: I1004 04:54:09.510673 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:54:09 crc kubenswrapper[4726]: E1004 04:54:09.513639 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:54:15 crc kubenswrapper[4726]: I1004 04:54:15.088047 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5b1149f2-4d8f-4b1e-b5ec-b13f9833c706/memcached/0.log" Oct 04 04:54:20 crc kubenswrapper[4726]: I1004 04:54:20.502262 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:54:20 crc kubenswrapper[4726]: E1004 04:54:20.503146 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:54:32 crc kubenswrapper[4726]: I1004 04:54:32.502418 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:54:32 crc kubenswrapper[4726]: E1004 04:54:32.503500 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:54:41 crc kubenswrapper[4726]: I1004 04:54:41.201347 4726 generic.go:334] "Generic (PLEG): container finished" podID="d49937fb-d320-4c40-8727-a0d224e719e2" containerID="ff253348e07af6b66ac29dd7f071f26b87427e8e3545f3008e1d11ce3bb5de36" exitCode=0 Oct 04 04:54:41 crc kubenswrapper[4726]: I1004 04:54:41.201452 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-tqq28" event={"ID":"d49937fb-d320-4c40-8727-a0d224e719e2","Type":"ContainerDied","Data":"ff253348e07af6b66ac29dd7f071f26b87427e8e3545f3008e1d11ce3bb5de36"} Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.687301 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.726852 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-tqq28"] Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.737290 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-tqq28"] Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.837656 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcrlr\" (UniqueName: \"kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr\") pod \"d49937fb-d320-4c40-8727-a0d224e719e2\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.837893 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host\") pod \"d49937fb-d320-4c40-8727-a0d224e719e2\" (UID: \"d49937fb-d320-4c40-8727-a0d224e719e2\") " Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.838042 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host" (OuterVolumeSpecName: "host") pod "d49937fb-d320-4c40-8727-a0d224e719e2" (UID: "d49937fb-d320-4c40-8727-a0d224e719e2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.838668 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d49937fb-d320-4c40-8727-a0d224e719e2-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.844937 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr" (OuterVolumeSpecName: "kube-api-access-wcrlr") pod "d49937fb-d320-4c40-8727-a0d224e719e2" (UID: "d49937fb-d320-4c40-8727-a0d224e719e2"). InnerVolumeSpecName "kube-api-access-wcrlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:42 crc kubenswrapper[4726]: I1004 04:54:42.940723 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcrlr\" (UniqueName: \"kubernetes.io/projected/d49937fb-d320-4c40-8727-a0d224e719e2-kube-api-access-wcrlr\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.220142 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca75e1e57fc8ace009322bcc4e7d4b04f52843968e745d09401a856aee5889af" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.220184 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-tqq28" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.899945 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lk69/crc-debug-q2gv2"] Oct 04 04:54:43 crc kubenswrapper[4726]: E1004 04:54:43.900644 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d49937fb-d320-4c40-8727-a0d224e719e2" containerName="container-00" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.900657 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="d49937fb-d320-4c40-8727-a0d224e719e2" containerName="container-00" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.900831 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="d49937fb-d320-4c40-8727-a0d224e719e2" containerName="container-00" Oct 04 04:54:43 crc kubenswrapper[4726]: I1004 04:54:43.901482 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.063299 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56czg\" (UniqueName: \"kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.063385 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.165749 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.165854 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.166208 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56czg\" (UniqueName: \"kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.186030 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56czg\" (UniqueName: \"kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg\") pod \"crc-debug-q2gv2\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.218868 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.504345 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:54:44 crc kubenswrapper[4726]: E1004 04:54:44.505030 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:54:44 crc kubenswrapper[4726]: I1004 04:54:44.515151 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d49937fb-d320-4c40-8727-a0d224e719e2" path="/var/lib/kubelet/pods/d49937fb-d320-4c40-8727-a0d224e719e2/volumes" Oct 04 04:54:45 crc kubenswrapper[4726]: I1004 04:54:45.240657 4726 generic.go:334] "Generic (PLEG): container finished" podID="af7c71a1-6467-49eb-b219-1e3929326a4a" containerID="e79feab32cde5cf8ee225c3799eb3f22d8e5e50164e7366f5324ea9f896a6c1b" exitCode=0 Oct 04 04:54:45 crc kubenswrapper[4726]: I1004 04:54:45.240725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" event={"ID":"af7c71a1-6467-49eb-b219-1e3929326a4a","Type":"ContainerDied","Data":"e79feab32cde5cf8ee225c3799eb3f22d8e5e50164e7366f5324ea9f896a6c1b"} Oct 04 04:54:45 crc kubenswrapper[4726]: I1004 04:54:45.241058 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" event={"ID":"af7c71a1-6467-49eb-b219-1e3929326a4a","Type":"ContainerStarted","Data":"7321bea13724145709e687de976d41628baa6607d0d9f9f450972cda584b5978"} Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.337393 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.503800 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host\") pod \"af7c71a1-6467-49eb-b219-1e3929326a4a\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.503895 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host" (OuterVolumeSpecName: "host") pod "af7c71a1-6467-49eb-b219-1e3929326a4a" (UID: "af7c71a1-6467-49eb-b219-1e3929326a4a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.503917 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56czg\" (UniqueName: \"kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg\") pod \"af7c71a1-6467-49eb-b219-1e3929326a4a\" (UID: \"af7c71a1-6467-49eb-b219-1e3929326a4a\") " Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.511393 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/af7c71a1-6467-49eb-b219-1e3929326a4a-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.574320 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg" (OuterVolumeSpecName: "kube-api-access-56czg") pod "af7c71a1-6467-49eb-b219-1e3929326a4a" (UID: "af7c71a1-6467-49eb-b219-1e3929326a4a"). InnerVolumeSpecName "kube-api-access-56czg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:46 crc kubenswrapper[4726]: I1004 04:54:46.613252 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56czg\" (UniqueName: \"kubernetes.io/projected/af7c71a1-6467-49eb-b219-1e3929326a4a-kube-api-access-56czg\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4726]: I1004 04:54:47.261005 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" event={"ID":"af7c71a1-6467-49eb-b219-1e3929326a4a","Type":"ContainerDied","Data":"7321bea13724145709e687de976d41628baa6607d0d9f9f450972cda584b5978"} Oct 04 04:54:47 crc kubenswrapper[4726]: I1004 04:54:47.261480 4726 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7321bea13724145709e687de976d41628baa6607d0d9f9f450972cda584b5978" Oct 04 04:54:47 crc kubenswrapper[4726]: I1004 04:54:47.261292 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-q2gv2" Oct 04 04:54:52 crc kubenswrapper[4726]: I1004 04:54:52.339174 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-q2gv2"] Oct 04 04:54:52 crc kubenswrapper[4726]: I1004 04:54:52.348214 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-q2gv2"] Oct 04 04:54:52 crc kubenswrapper[4726]: I1004 04:54:52.514924 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af7c71a1-6467-49eb-b219-1e3929326a4a" path="/var/lib/kubelet/pods/af7c71a1-6467-49eb-b219-1e3929326a4a/volumes" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.505994 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7lk69/crc-debug-gc5tc"] Oct 04 04:54:53 crc kubenswrapper[4726]: E1004 04:54:53.507153 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7c71a1-6467-49eb-b219-1e3929326a4a" containerName="container-00" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.507253 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7c71a1-6467-49eb-b219-1e3929326a4a" containerName="container-00" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.507549 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7c71a1-6467-49eb-b219-1e3929326a4a" containerName="container-00" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.508311 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.623621 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pcnk\" (UniqueName: \"kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.624378 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.726063 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pcnk\" (UniqueName: \"kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.726207 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.726373 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.747574 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pcnk\" (UniqueName: \"kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk\") pod \"crc-debug-gc5tc\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:53 crc kubenswrapper[4726]: I1004 04:54:53.827885 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.024409 4726 scope.go:117] "RemoveContainer" containerID="b29db3b88ab435dd5b7cdc210cc024c5689fceca1b408eb03ce2c2e349959737" Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.330869 4726 generic.go:334] "Generic (PLEG): container finished" podID="fbf3b46b-3521-49ae-8a5e-0debeb585cba" containerID="3c7307b96e270526b6ec7489a9b3e214e09845c90b13d0fdffa4992a9cb82adf" exitCode=0 Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.330924 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" event={"ID":"fbf3b46b-3521-49ae-8a5e-0debeb585cba","Type":"ContainerDied","Data":"3c7307b96e270526b6ec7489a9b3e214e09845c90b13d0fdffa4992a9cb82adf"} Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.330963 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" event={"ID":"fbf3b46b-3521-49ae-8a5e-0debeb585cba","Type":"ContainerStarted","Data":"67aa2a7a2be64377f9a3949ec05fe79600c849fd316d8557b4d7e4e0b42cae97"} Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.368049 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-gc5tc"] Oct 04 04:54:54 crc kubenswrapper[4726]: I1004 04:54:54.376233 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lk69/crc-debug-gc5tc"] Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.451831 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.557328 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pcnk\" (UniqueName: \"kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk\") pod \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.557506 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host\") pod \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\" (UID: \"fbf3b46b-3521-49ae-8a5e-0debeb585cba\") " Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.557794 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host" (OuterVolumeSpecName: "host") pod "fbf3b46b-3521-49ae-8a5e-0debeb585cba" (UID: "fbf3b46b-3521-49ae-8a5e-0debeb585cba"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.558337 4726 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fbf3b46b-3521-49ae-8a5e-0debeb585cba-host\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.575248 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk" (OuterVolumeSpecName: "kube-api-access-7pcnk") pod "fbf3b46b-3521-49ae-8a5e-0debeb585cba" (UID: "fbf3b46b-3521-49ae-8a5e-0debeb585cba"). InnerVolumeSpecName "kube-api-access-7pcnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.660421 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pcnk\" (UniqueName: \"kubernetes.io/projected/fbf3b46b-3521-49ae-8a5e-0debeb585cba-kube-api-access-7pcnk\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:55 crc kubenswrapper[4726]: I1004 04:54:55.971887 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.160738 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.163251 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.191294 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.350226 4726 scope.go:117] "RemoveContainer" containerID="3c7307b96e270526b6ec7489a9b3e214e09845c90b13d0fdffa4992a9cb82adf" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.350306 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/crc-debug-gc5tc" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.361645 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/pull/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.403116 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/util/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.413685 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1606e2ec47c0e272775e8116fb06ab975432a286eb57580e4b04364177r582x_12b6c390-7847-4e46-bdd6-6f96a1780f18/extract/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.523799 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf3b46b-3521-49ae-8a5e-0debeb585cba" path="/var/lib/kubelet/pods/fbf3b46b-3521-49ae-8a5e-0debeb585cba/volumes" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.557609 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-mfsl5_1e47df33-2621-4cfe-9c83-4d4bd1d616aa/kube-rbac-proxy/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.625235 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-tzr27_f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3/kube-rbac-proxy/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.676504 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-mfsl5_1e47df33-2621-4cfe-9c83-4d4bd1d616aa/manager/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.811905 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-tzr27_f8f53e59-6d0a-4dd0-b63f-b12f73cc3fd3/manager/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.903980 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-gtshn_a520bf75-9223-4209-87d4-248ce3e0d190/kube-rbac-proxy/0.log" Oct 04 04:54:56 crc kubenswrapper[4726]: I1004 04:54:56.915895 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-gtshn_a520bf75-9223-4209-87d4-248ce3e0d190/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.057936 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-cjx6d_79a719e7-5bce-424a-b837-ab0bca45936e/kube-rbac-proxy/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.186614 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-cjx6d_79a719e7-5bce-424a-b837-ab0bca45936e/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.217946 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-lqs6l_ac49881b-2d45-4ba0-8b60-8461a2d5f35b/kube-rbac-proxy/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.269126 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-lqs6l_ac49881b-2d45-4ba0-8b60-8461a2d5f35b/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.397473 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-fvw5p_5e5f2785-db94-4fdc-9cb2-10b1ce940694/kube-rbac-proxy/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.433924 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-fvw5p_5e5f2785-db94-4fdc-9cb2-10b1ce940694/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.543261 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7ddb876846-kgv8l_ea261e29-40c5-4a26-a02f-8a09e0c7320d/kube-rbac-proxy/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.649300 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-v72j6_9f398c24-0435-49f5-915c-a54efeadea0c/kube-rbac-proxy/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.772804 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-v72j6_9f398c24-0435-49f5-915c-a54efeadea0c/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.779987 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7ddb876846-kgv8l_ea261e29-40c5-4a26-a02f-8a09e0c7320d/manager/0.log" Oct 04 04:54:57 crc kubenswrapper[4726]: I1004 04:54:57.861608 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-wwdj2_488fc85d-f45f-4d8d-93e2-b58ba9306a76/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.027369 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-wwdj2_488fc85d-f45f-4d8d-93e2-b58ba9306a76/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.063853 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-6dpjp_6508ae63-4f75-496f-b7b6-5dcd8a0cb008/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.138208 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-6dpjp_6508ae63-4f75-496f-b7b6-5dcd8a0cb008/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.198001 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-s67gj_d55e66e5-0c46-4df2-909c-fc739a26c5cc/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.292616 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-s67gj_d55e66e5-0c46-4df2-909c-fc739a26c5cc/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.394778 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lqsxr_d59f6f6b-419b-4ce3-adb6-11381aed6a5d/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.444090 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-lqsxr_d59f6f6b-419b-4ce3-adb6-11381aed6a5d/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.506155 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:54:58 crc kubenswrapper[4726]: E1004 04:54:58.506449 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.560596 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mnndm_05b7da99-96e2-40be-8fc1-f847b8e7a643/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.647541 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mnndm_05b7da99-96e2-40be-8fc1-f847b8e7a643/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.746216 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-zkgst_cadfc4b1-9327-4a55-8544-44422572319c/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.750299 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-zkgst_cadfc4b1-9327-4a55-8544-44422572319c/manager/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.868166 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt_6b245f13-4e4b-4a41-b148-482d581e2c5c/kube-rbac-proxy/0.log" Oct 04 04:54:58 crc kubenswrapper[4726]: I1004 04:54:58.915635 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c8mwjt_6b245f13-4e4b-4a41-b148-482d581e2c5c/manager/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.053933 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-ff7c6f545-hgqmd_aaa6fd43-d6ed-4d86-9827-27d2529499e3/kube-rbac-proxy/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.160451 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5499555f6-blzgx_0bd87ab8-94f6-41a4-b76c-7774cac7b625/kube-rbac-proxy/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.433851 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fzc75_40e39198-082e-4c71-b67c-6d845e45f782/registry-server/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.485885 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5499555f6-blzgx_0bd87ab8-94f6-41a4-b76c-7774cac7b625/operator/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.702461 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-8pnpc_e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828/kube-rbac-proxy/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.787326 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-8pnpc_e9f7cf26-4f1f-4d3a-8016-cbe4c25c3828/manager/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.867388 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-lsxtw_efa34c5f-e531-4319-b5fb-7497c4d026f0/kube-rbac-proxy/0.log" Oct 04 04:54:59 crc kubenswrapper[4726]: I1004 04:54:59.995468 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-lsxtw_efa34c5f-e531-4319-b5fb-7497c4d026f0/manager/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.097279 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-pkjd7_962ba8ee-fbcf-4ef5-b740-bd4a782d0881/operator/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.221157 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hv9lz_5cb54ad2-85db-4291-87dc-4f0b9ea688c0/kube-rbac-proxy/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.256569 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-ff7c6f545-hgqmd_aaa6fd43-d6ed-4d86-9827-27d2529499e3/manager/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.337530 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-hv9lz_5cb54ad2-85db-4291-87dc-4f0b9ea688c0/manager/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.418421 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-2j9rb_42949778-256b-4809-b2ce-d633428e649c/kube-rbac-proxy/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.454998 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-2j9rb_42949778-256b-4809-b2ce-d633428e649c/manager/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.522448 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4v7k8_392e7503-f32e-4e8f-ac07-d75844253ea1/kube-rbac-proxy/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.580228 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-4v7k8_392e7503-f32e-4e8f-ac07-d75844253ea1/manager/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.678047 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-kn2pc_262abd5f-25d1-488f-b864-f9cc308a68f4/kube-rbac-proxy/0.log" Oct 04 04:55:00 crc kubenswrapper[4726]: I1004 04:55:00.679585 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-kn2pc_262abd5f-25d1-488f-b864-f9cc308a68f4/manager/0.log" Oct 04 04:55:12 crc kubenswrapper[4726]: I1004 04:55:12.502444 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:55:12 crc kubenswrapper[4726]: E1004 04:55:12.503428 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:55:17 crc kubenswrapper[4726]: I1004 04:55:17.500325 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qstzg_fa73eb1e-f78b-41f1-9be1-aeae979d66f9/control-plane-machine-set-operator/0.log" Oct 04 04:55:17 crc kubenswrapper[4726]: I1004 04:55:17.699340 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ffl97_7386c233-34f0-4eef-87d0-ac2b7b145897/kube-rbac-proxy/0.log" Oct 04 04:55:17 crc kubenswrapper[4726]: I1004 04:55:17.762757 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ffl97_7386c233-34f0-4eef-87d0-ac2b7b145897/machine-api-operator/0.log" Oct 04 04:55:25 crc kubenswrapper[4726]: I1004 04:55:25.502705 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:55:25 crc kubenswrapper[4726]: E1004 04:55:25.503575 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:55:30 crc kubenswrapper[4726]: I1004 04:55:30.055133 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vbkwl_53efe0a4-7f9a-4541-9677-12e0dcec292e/cert-manager-controller/0.log" Oct 04 04:55:30 crc kubenswrapper[4726]: I1004 04:55:30.175707 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hkk79_3628e5db-697e-495c-815f-c08b5ccd03da/cert-manager-cainjector/0.log" Oct 04 04:55:30 crc kubenswrapper[4726]: I1004 04:55:30.283539 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-nkt2j_5e7b44b7-601a-4d34-a2f5-13662f3ef3cd/cert-manager-webhook/0.log" Oct 04 04:55:38 crc kubenswrapper[4726]: I1004 04:55:38.502882 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:55:38 crc kubenswrapper[4726]: E1004 04:55:38.503815 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:55:43 crc kubenswrapper[4726]: I1004 04:55:43.626508 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-2tqv5_0afe634c-ed15-4abf-998a-5929ab4c94c3/nmstate-console-plugin/0.log" Oct 04 04:55:43 crc kubenswrapper[4726]: I1004 04:55:43.788260 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jx5pc_b0606264-b345-49e2-9965-806462686e84/nmstate-handler/0.log" Oct 04 04:55:43 crc kubenswrapper[4726]: I1004 04:55:43.829268 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8n4qb_10fdef1b-c247-4b30-82b9-cde45b292b17/nmstate-metrics/0.log" Oct 04 04:55:43 crc kubenswrapper[4726]: I1004 04:55:43.848935 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8n4qb_10fdef1b-c247-4b30-82b9-cde45b292b17/kube-rbac-proxy/0.log" Oct 04 04:55:43 crc kubenswrapper[4726]: I1004 04:55:43.991030 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-s4p69_a5ee8c2e-8f69-4301-8aab-c5af01402c8a/nmstate-operator/0.log" Oct 04 04:55:44 crc kubenswrapper[4726]: I1004 04:55:44.044045 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-mxq57_b2974204-71e0-4cd1-859b-0aef97909ee9/nmstate-webhook/0.log" Oct 04 04:55:49 crc kubenswrapper[4726]: I1004 04:55:49.503236 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:55:49 crc kubenswrapper[4726]: E1004 04:55:49.504052 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:55:57 crc kubenswrapper[4726]: I1004 04:55:57.759997 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tv4tz_eca76b37-13e9-4101-8486-7d009410a7bb/kube-rbac-proxy/0.log" Oct 04 04:55:57 crc kubenswrapper[4726]: I1004 04:55:57.811569 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-tv4tz_eca76b37-13e9-4101-8486-7d009410a7bb/controller/0.log" Oct 04 04:55:57 crc kubenswrapper[4726]: I1004 04:55:57.946554 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-76nkf_555989d8-4bc6-4750-a532-f71f09933d71/frr-k8s-webhook-server/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.036633 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.180669 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.227012 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.258502 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.261327 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.386833 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.444618 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.487051 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.508410 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.681618 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-metrics/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.695636 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-reloader/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.837401 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/cp-frr-files/0.log" Oct 04 04:55:58 crc kubenswrapper[4726]: I1004 04:55:58.868075 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/controller/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.049428 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/kube-rbac-proxy-frr/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.086901 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/frr-metrics/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.158618 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/kube-rbac-proxy/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.326091 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/reloader/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.471628 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-655cdc4dc4-v2hjh_898913f8-1833-4c8f-8764-32c656d6caab/manager/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.650283 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66d48bf4d6-nkg69_898ba0b6-6440-45be-8067-6719ead41cba/webhook-server/0.log" Oct 04 04:55:59 crc kubenswrapper[4726]: I1004 04:55:59.782819 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89wr2_ca2e3d23-2c1d-4b4a-85d6-c280ce107167/kube-rbac-proxy/0.log" Oct 04 04:56:00 crc kubenswrapper[4726]: I1004 04:56:00.244884 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-89wr2_ca2e3d23-2c1d-4b4a-85d6-c280ce107167/speaker/0.log" Oct 04 04:56:00 crc kubenswrapper[4726]: I1004 04:56:00.433955 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zktqf_98ad93ad-5869-4d8d-8074-c39600f257b1/frr/0.log" Oct 04 04:56:02 crc kubenswrapper[4726]: I1004 04:56:02.502907 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:56:02 crc kubenswrapper[4726]: E1004 04:56:02.503522 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:56:12 crc kubenswrapper[4726]: I1004 04:56:12.941059 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.114787 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.134584 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.162129 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.340462 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/pull/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.381475 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/util/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.397498 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2rcj9p_decbcd72-7ac4-4806-9b47-9611634efc23/extract/0.log" Oct 04 04:56:13 crc kubenswrapper[4726]: I1004 04:56:13.503068 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:56:13 crc kubenswrapper[4726]: E1004 04:56:13.503340 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.076067 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.236741 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.254221 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.280310 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.479404 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-utilities/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.518611 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/extract-content/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.780684 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.960349 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:56:14 crc kubenswrapper[4726]: I1004 04:56:14.999474 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.011327 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ptxbw_afe2d9b4-beff-42d2-b599-441fcb58ddbd/registry-server/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.045018 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.233973 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-content/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.238847 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/extract-utilities/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.441439 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:56:15 crc kubenswrapper[4726]: I1004 04:56:15.840468 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-65mdl_9d6ba11a-3030-4fb4-a3d9-f46bf8c72a4a/registry-server/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.133302 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.219322 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.236862 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.392189 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/extract/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.419861 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/util/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.452089 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cjttz5_afad42c8-8dff-4c05-a9b8-08f750cbedec/pull/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.454377 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mqgpt_f0575ad7-500b-4c8e-9bbd-843e6bc0a5e8/marketplace-operator/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.631289 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.756699 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.775646 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.790472 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.969937 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-content/0.log" Oct 04 04:56:16 crc kubenswrapper[4726]: I1004 04:56:16.970137 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/extract-utilities/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.025259 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.164405 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7skgq_b10c732f-90f0-413e-8efe-8ae89ca2ea39/registry-server/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.232950 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.240668 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.251199 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.412599 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-utilities/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.436668 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/extract-content/0.log" Oct 04 04:56:17 crc kubenswrapper[4726]: I1004 04:56:17.896232 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-dnxfd_c45c00c6-5a4c-4c73-935b-185cf2759ea7/registry-server/0.log" Oct 04 04:56:24 crc kubenswrapper[4726]: I1004 04:56:24.502791 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:56:24 crc kubenswrapper[4726]: E1004 04:56:24.504439 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:56:39 crc kubenswrapper[4726]: I1004 04:56:39.503349 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:56:39 crc kubenswrapper[4726]: E1004 04:56:39.504208 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:56:51 crc kubenswrapper[4726]: I1004 04:56:51.503851 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:56:51 crc kubenswrapper[4726]: E1004 04:56:51.506296 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:57:05 crc kubenswrapper[4726]: I1004 04:57:05.502812 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:57:05 crc kubenswrapper[4726]: E1004 04:57:05.504524 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:57:18 crc kubenswrapper[4726]: I1004 04:57:18.508007 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:57:18 crc kubenswrapper[4726]: E1004 04:57:18.508848 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:57:29 crc kubenswrapper[4726]: I1004 04:57:29.503030 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:57:29 crc kubenswrapper[4726]: E1004 04:57:29.503856 4726 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9777l_openshift-machine-config-operator(e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1)\"" pod="openshift-machine-config-operator/machine-config-daemon-9777l" podUID="e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.358212 4726 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:33 crc kubenswrapper[4726]: E1004 04:57:33.359510 4726 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf3b46b-3521-49ae-8a5e-0debeb585cba" containerName="container-00" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.359531 4726 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf3b46b-3521-49ae-8a5e-0debeb585cba" containerName="container-00" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.359822 4726 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf3b46b-3521-49ae-8a5e-0debeb585cba" containerName="container-00" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.363474 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.377218 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.437243 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.437524 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.437567 4726 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq8nv\" (UniqueName: \"kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.539757 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.539817 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.539878 4726 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq8nv\" (UniqueName: \"kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.540454 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.541138 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.564647 4726 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq8nv\" (UniqueName: \"kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv\") pod \"redhat-operators-7h22g\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:33 crc kubenswrapper[4726]: I1004 04:57:33.695029 4726 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:34 crc kubenswrapper[4726]: I1004 04:57:34.181630 4726 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:34 crc kubenswrapper[4726]: I1004 04:57:34.775509 4726 generic.go:334] "Generic (PLEG): container finished" podID="02f755ce-d9b0-4dc1-83d6-944936ace5d8" containerID="46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40" exitCode=0 Oct 04 04:57:34 crc kubenswrapper[4726]: I1004 04:57:34.775755 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerDied","Data":"46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40"} Oct 04 04:57:34 crc kubenswrapper[4726]: I1004 04:57:34.777402 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerStarted","Data":"facaa3fc13fc32cdadf1dd845c518238d4eceb913881cd8a98def158ce66fe7d"} Oct 04 04:57:34 crc kubenswrapper[4726]: I1004 04:57:34.777317 4726 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:57:36 crc kubenswrapper[4726]: I1004 04:57:36.798745 4726 generic.go:334] "Generic (PLEG): container finished" podID="02f755ce-d9b0-4dc1-83d6-944936ace5d8" containerID="6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c" exitCode=0 Oct 04 04:57:36 crc kubenswrapper[4726]: I1004 04:57:36.798802 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerDied","Data":"6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c"} Oct 04 04:57:38 crc kubenswrapper[4726]: I1004 04:57:38.826725 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerStarted","Data":"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56"} Oct 04 04:57:38 crc kubenswrapper[4726]: I1004 04:57:38.850339 4726 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7h22g" podStartSLOduration=2.391132811 podStartE2EDuration="5.850321359s" podCreationTimestamp="2025-10-04 04:57:33 +0000 UTC" firstStartedPulling="2025-10-04 04:57:34.777040586 +0000 UTC m=+4628.951663799" lastFinishedPulling="2025-10-04 04:57:38.236229114 +0000 UTC m=+4632.410852347" observedRunningTime="2025-10-04 04:57:38.847075274 +0000 UTC m=+4633.021698487" watchObservedRunningTime="2025-10-04 04:57:38.850321359 +0000 UTC m=+4633.024944572" Oct 04 04:57:41 crc kubenswrapper[4726]: I1004 04:57:41.502916 4726 scope.go:117] "RemoveContainer" containerID="1f96e67061977f5b6ff7e2fea44a5d7516472cc52dfc5684ea3149004564b6ec" Oct 04 04:57:42 crc kubenswrapper[4726]: I1004 04:57:42.865954 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9777l" event={"ID":"e3161ea0-64d9-4f8b-8ef5-6721e3a58ff1","Type":"ContainerStarted","Data":"bd7924e343b1ee784bcad286c4bf5cb040e02ed1f2fca53fec63060bac28a89c"} Oct 04 04:57:43 crc kubenswrapper[4726]: I1004 04:57:43.695721 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:43 crc kubenswrapper[4726]: I1004 04:57:43.696100 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:43 crc kubenswrapper[4726]: I1004 04:57:43.767009 4726 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:43 crc kubenswrapper[4726]: I1004 04:57:43.934054 4726 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:44 crc kubenswrapper[4726]: I1004 04:57:44.008910 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:45 crc kubenswrapper[4726]: I1004 04:57:45.904261 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7h22g" podUID="02f755ce-d9b0-4dc1-83d6-944936ace5d8" containerName="registry-server" containerID="cri-o://ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56" gracePeriod=2 Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.856075 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.904148 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq8nv\" (UniqueName: \"kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv\") pod \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.904224 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities\") pod \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.904271 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content\") pod \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\" (UID: \"02f755ce-d9b0-4dc1-83d6-944936ace5d8\") " Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.905414 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities" (OuterVolumeSpecName: "utilities") pod "02f755ce-d9b0-4dc1-83d6-944936ace5d8" (UID: "02f755ce-d9b0-4dc1-83d6-944936ace5d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.911965 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv" (OuterVolumeSpecName: "kube-api-access-kq8nv") pod "02f755ce-d9b0-4dc1-83d6-944936ace5d8" (UID: "02f755ce-d9b0-4dc1-83d6-944936ace5d8"). InnerVolumeSpecName "kube-api-access-kq8nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.916619 4726 generic.go:334] "Generic (PLEG): container finished" podID="02f755ce-d9b0-4dc1-83d6-944936ace5d8" containerID="ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56" exitCode=0 Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.916694 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerDied","Data":"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56"} Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.916726 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h22g" event={"ID":"02f755ce-d9b0-4dc1-83d6-944936ace5d8","Type":"ContainerDied","Data":"facaa3fc13fc32cdadf1dd845c518238d4eceb913881cd8a98def158ce66fe7d"} Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.916749 4726 scope.go:117] "RemoveContainer" containerID="ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.916902 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h22g" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.968574 4726 scope.go:117] "RemoveContainer" containerID="6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c" Oct 04 04:57:46 crc kubenswrapper[4726]: I1004 04:57:46.990858 4726 scope.go:117] "RemoveContainer" containerID="46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.000263 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02f755ce-d9b0-4dc1-83d6-944936ace5d8" (UID: "02f755ce-d9b0-4dc1-83d6-944936ace5d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.007997 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq8nv\" (UniqueName: \"kubernetes.io/projected/02f755ce-d9b0-4dc1-83d6-944936ace5d8-kube-api-access-kq8nv\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.008041 4726 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.008051 4726 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02f755ce-d9b0-4dc1-83d6-944936ace5d8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.056753 4726 scope.go:117] "RemoveContainer" containerID="ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56" Oct 04 04:57:47 crc kubenswrapper[4726]: E1004 04:57:47.057655 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56\": container with ID starting with ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56 not found: ID does not exist" containerID="ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.057696 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56"} err="failed to get container status \"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56\": rpc error: code = NotFound desc = could not find container \"ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56\": container with ID starting with ab5bf20f0e1b72120245ca4c90a5876f8769f2923fb62dc3bd508d58a2664f56 not found: ID does not exist" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.057724 4726 scope.go:117] "RemoveContainer" containerID="6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c" Oct 04 04:57:47 crc kubenswrapper[4726]: E1004 04:57:47.058027 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c\": container with ID starting with 6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c not found: ID does not exist" containerID="6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.058070 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c"} err="failed to get container status \"6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c\": rpc error: code = NotFound desc = could not find container \"6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c\": container with ID starting with 6dadf8737590cfffebba03f60fa2674cb2ca447b86a673509a33dab0f5f2792c not found: ID does not exist" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.058136 4726 scope.go:117] "RemoveContainer" containerID="46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40" Oct 04 04:57:47 crc kubenswrapper[4726]: E1004 04:57:47.058416 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40\": container with ID starting with 46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40 not found: ID does not exist" containerID="46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.058448 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40"} err="failed to get container status \"46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40\": rpc error: code = NotFound desc = could not find container \"46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40\": container with ID starting with 46592c0957bbace945fd99307964573af0ca9cbeb3270b0a1cf5fd7395d5ef40 not found: ID does not exist" Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.254814 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:47 crc kubenswrapper[4726]: I1004 04:57:47.261831 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7h22g"] Oct 04 04:57:48 crc kubenswrapper[4726]: I1004 04:57:48.517807 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f755ce-d9b0-4dc1-83d6-944936ace5d8" path="/var/lib/kubelet/pods/02f755ce-d9b0-4dc1-83d6-944936ace5d8/volumes" Oct 04 04:58:17 crc kubenswrapper[4726]: I1004 04:58:17.189629 4726 generic.go:334] "Generic (PLEG): container finished" podID="97accf81-66a4-49ff-8228-0505dc1bfd94" containerID="a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1" exitCode=0 Oct 04 04:58:17 crc kubenswrapper[4726]: I1004 04:58:17.189702 4726 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7lk69/must-gather-lwbct" event={"ID":"97accf81-66a4-49ff-8228-0505dc1bfd94","Type":"ContainerDied","Data":"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1"} Oct 04 04:58:17 crc kubenswrapper[4726]: I1004 04:58:17.190853 4726 scope.go:117] "RemoveContainer" containerID="a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1" Oct 04 04:58:18 crc kubenswrapper[4726]: I1004 04:58:18.010537 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lk69_must-gather-lwbct_97accf81-66a4-49ff-8228-0505dc1bfd94/gather/0.log" Oct 04 04:58:29 crc kubenswrapper[4726]: I1004 04:58:29.737728 4726 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7lk69/must-gather-lwbct"] Oct 04 04:58:29 crc kubenswrapper[4726]: I1004 04:58:29.738726 4726 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7lk69/must-gather-lwbct" podUID="97accf81-66a4-49ff-8228-0505dc1bfd94" containerName="copy" containerID="cri-o://86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5" gracePeriod=2 Oct 04 04:58:29 crc kubenswrapper[4726]: I1004 04:58:29.745150 4726 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7lk69/must-gather-lwbct"] Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.266392 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lk69_must-gather-lwbct_97accf81-66a4-49ff-8228-0505dc1bfd94/copy/0.log" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.267085 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.315587 4726 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7lk69_must-gather-lwbct_97accf81-66a4-49ff-8228-0505dc1bfd94/copy/0.log" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.316178 4726 generic.go:334] "Generic (PLEG): container finished" podID="97accf81-66a4-49ff-8228-0505dc1bfd94" containerID="86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5" exitCode=143 Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.316224 4726 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7lk69/must-gather-lwbct" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.316252 4726 scope.go:117] "RemoveContainer" containerID="86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.337683 4726 scope.go:117] "RemoveContainer" containerID="a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.398162 4726 scope.go:117] "RemoveContainer" containerID="86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5" Oct 04 04:58:30 crc kubenswrapper[4726]: E1004 04:58:30.398750 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5\": container with ID starting with 86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5 not found: ID does not exist" containerID="86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.398811 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5"} err="failed to get container status \"86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5\": rpc error: code = NotFound desc = could not find container \"86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5\": container with ID starting with 86dd37614de8a25c7b4eb158138f44305de60e199f03d1f435ec7d08e04b34b5 not found: ID does not exist" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.398844 4726 scope.go:117] "RemoveContainer" containerID="a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1" Oct 04 04:58:30 crc kubenswrapper[4726]: E1004 04:58:30.399298 4726 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1\": container with ID starting with a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1 not found: ID does not exist" containerID="a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.399354 4726 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1"} err="failed to get container status \"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1\": rpc error: code = NotFound desc = could not find container \"a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1\": container with ID starting with a8711deaecafc143cd94e0b33b38cbe4d98678745844333cc52158749056abf1 not found: ID does not exist" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.439747 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnrf6\" (UniqueName: \"kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6\") pod \"97accf81-66a4-49ff-8228-0505dc1bfd94\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.440013 4726 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output\") pod \"97accf81-66a4-49ff-8228-0505dc1bfd94\" (UID: \"97accf81-66a4-49ff-8228-0505dc1bfd94\") " Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.445602 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6" (OuterVolumeSpecName: "kube-api-access-tnrf6") pod "97accf81-66a4-49ff-8228-0505dc1bfd94" (UID: "97accf81-66a4-49ff-8228-0505dc1bfd94"). InnerVolumeSpecName "kube-api-access-tnrf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.546616 4726 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnrf6\" (UniqueName: \"kubernetes.io/projected/97accf81-66a4-49ff-8228-0505dc1bfd94-kube-api-access-tnrf6\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.633873 4726 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "97accf81-66a4-49ff-8228-0505dc1bfd94" (UID: "97accf81-66a4-49ff-8228-0505dc1bfd94"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:30 crc kubenswrapper[4726]: I1004 04:58:30.648692 4726 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/97accf81-66a4-49ff-8228-0505dc1bfd94-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:32 crc kubenswrapper[4726]: I1004 04:58:32.513419 4726 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97accf81-66a4-49ff-8228-0505dc1bfd94" path="/var/lib/kubelet/pods/97accf81-66a4-49ff-8228-0505dc1bfd94/volumes" Oct 04 04:58:54 crc kubenswrapper[4726]: I1004 04:58:54.396227 4726 scope.go:117] "RemoveContainer" containerID="ff253348e07af6b66ac29dd7f071f26b87427e8e3545f3008e1d11ce3bb5de36" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070124713024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070124713017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070113135016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070113136015452 5ustar corecore